[ 445.466521] env[61768]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61768) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 445.466925] env[61768]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61768) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 445.466986] env[61768]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61768) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 445.467297] env[61768]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 445.562573] env[61768]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61768) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 445.572502] env[61768]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61768) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 446.173227] env[61768]: INFO nova.virt.driver [None req-476b1adf-e555-44fe-a310-91ab2b75b897 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 446.244661] env[61768]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 446.244842] env[61768]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 446.244915] env[61768]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61768) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 449.314927] env[61768]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-1115f15c-c24b-4af3-81c4-c5e3daed33a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.330656] env[61768]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61768) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 449.330806] env[61768]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-b3c6e228-5618-4805-8b47-0eeaededb145 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.365943] env[61768]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 0dfc7. [ 449.366157] env[61768]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.121s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 449.366641] env[61768]: INFO nova.virt.vmwareapi.driver [None req-476b1adf-e555-44fe-a310-91ab2b75b897 None None] VMware vCenter version: 7.0.3 [ 449.370081] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370cd65e-fb1e-4887-b997-6c67c913e189 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.387236] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1386a00-0a83-4032-8c58-d24c9bbf197a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.392899] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3850e3c-0f64-42db-b225-d873258a66a2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.399210] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb626d0d-ada4-4bdf-9367-7672504a1176 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.411841] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1339d91d-5759-4059-a117-5cc31eb0d501 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.417528] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-430836c3-9ca6-48c1-830e-819db060e491 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.446613] env[61768]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-173d6c62-615d-4d1f-ba93-d3133c290a4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 449.451327] env[61768]: DEBUG nova.virt.vmwareapi.driver [None req-476b1adf-e555-44fe-a310-91ab2b75b897 None None] Extension org.openstack.compute already exists. {{(pid=61768) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 449.453972] env[61768]: INFO nova.compute.provider_config [None req-476b1adf-e555-44fe-a310-91ab2b75b897 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 449.957875] env[61768]: DEBUG nova.context [None req-476b1adf-e555-44fe-a310-91ab2b75b897 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),459ba036-6e22-4223-a3b3-0c8520fb08ab(cell1) {{(pid=61768) load_cells /opt/stack/nova/nova/context.py:464}} [ 449.959918] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 449.960163] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 449.960837] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 449.961287] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Acquiring lock "459ba036-6e22-4223-a3b3-0c8520fb08ab" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 449.961481] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Lock "459ba036-6e22-4223-a3b3-0c8520fb08ab" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 449.962532] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Lock "459ba036-6e22-4223-a3b3-0c8520fb08ab" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 449.982835] env[61768]: INFO dbcounter [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Registered counter for database nova_cell0 [ 449.990894] env[61768]: INFO dbcounter [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Registered counter for database nova_cell1 [ 449.994304] env[61768]: DEBUG oslo_db.sqlalchemy.engines [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61768) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 449.994656] env[61768]: DEBUG oslo_db.sqlalchemy.engines [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61768) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 449.999367] env[61768]: ERROR nova.db.main.api [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 449.999367] env[61768]: result = function(*args, **kwargs) [ 449.999367] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 449.999367] env[61768]: return func(*args, **kwargs) [ 449.999367] env[61768]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 449.999367] env[61768]: result = fn(*args, **kwargs) [ 449.999367] env[61768]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 449.999367] env[61768]: return f(*args, **kwargs) [ 449.999367] env[61768]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 449.999367] env[61768]: return db.service_get_minimum_version(context, binaries) [ 449.999367] env[61768]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 449.999367] env[61768]: _check_db_access() [ 449.999367] env[61768]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 449.999367] env[61768]: stacktrace = ''.join(traceback.format_stack()) [ 449.999367] env[61768]: [ 450.000164] env[61768]: ERROR nova.db.main.api [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 450.000164] env[61768]: result = function(*args, **kwargs) [ 450.000164] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 450.000164] env[61768]: return func(*args, **kwargs) [ 450.000164] env[61768]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 450.000164] env[61768]: result = fn(*args, **kwargs) [ 450.000164] env[61768]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 450.000164] env[61768]: return f(*args, **kwargs) [ 450.000164] env[61768]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 450.000164] env[61768]: return db.service_get_minimum_version(context, binaries) [ 450.000164] env[61768]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 450.000164] env[61768]: _check_db_access() [ 450.000164] env[61768]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 450.000164] env[61768]: stacktrace = ''.join(traceback.format_stack()) [ 450.000164] env[61768]: [ 450.000734] env[61768]: WARNING nova.objects.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Failed to get minimum service version for cell 459ba036-6e22-4223-a3b3-0c8520fb08ab [ 450.000734] env[61768]: WARNING nova.objects.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 450.001089] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Acquiring lock "singleton_lock" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 450.001255] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Acquired lock "singleton_lock" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 450.001496] env[61768]: DEBUG oslo_concurrency.lockutils [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Releasing lock "singleton_lock" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 450.001906] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Full set of CONF: {{(pid=61768) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 450.002077] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ******************************************************************************** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 450.002211] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Configuration options gathered from: {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 450.002349] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 450.002544] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 450.002672] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ================================================================================ {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 450.002880] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] allow_resize_to_same_host = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.003061] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] arq_binding_timeout = 300 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.003197] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] backdoor_port = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.003323] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] backdoor_socket = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.003484] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] block_device_allocate_retries = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.003644] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] block_device_allocate_retries_interval = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.003878] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cert = self.pem {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.004095] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.004277] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute_monitors = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.004449] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] config_dir = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.004622] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] config_drive_format = iso9660 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.004756] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.004939] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] config_source = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.005149] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] console_host = devstack {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.005320] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] control_exchange = nova {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.005479] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cpu_allocation_ratio = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.005642] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] daemon = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.005812] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] debug = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.005974] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] default_access_ip_network_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.006153] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] default_availability_zone = nova {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.006312] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] default_ephemeral_format = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.006472] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] default_green_pool_size = 1000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.006703] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.006868] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] default_schedule_zone = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.007038] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] disk_allocation_ratio = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.007205] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] enable_new_services = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.007384] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] enabled_apis = ['osapi_compute'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.007549] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] enabled_ssl_apis = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.007712] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] flat_injected = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.007872] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] force_config_drive = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.008040] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] force_raw_images = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.008211] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] graceful_shutdown_timeout = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.008371] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] heal_instance_info_cache_interval = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.008580] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] host = cpu-1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.008756] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.008919] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] initial_disk_allocation_ratio = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.009093] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] initial_ram_allocation_ratio = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.009319] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.009486] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instance_build_timeout = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.009646] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instance_delete_interval = 300 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.009815] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instance_format = [instance: %(uuid)s] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.009983] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instance_name_template = instance-%08x {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.010163] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instance_usage_audit = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.010336] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instance_usage_audit_period = month {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.010503] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.010669] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] instances_path = /opt/stack/data/nova/instances {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.010835] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] internal_service_availability_zone = internal {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.010997] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] key = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.011170] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] live_migration_retry_count = 30 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.011337] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_color = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.011505] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_config_append = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.011672] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.011926] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_dir = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.012139] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.012278] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_options = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.012445] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_rotate_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.012615] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_rotate_interval_type = days {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.012781] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] log_rotation_type = none {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.012910] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.013048] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.013222] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.013387] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.013515] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.013683] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] long_rpc_timeout = 1800 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.013915] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] max_concurrent_builds = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.014102] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] max_concurrent_live_migrations = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.014269] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] max_concurrent_snapshots = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.014434] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] max_local_block_devices = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.014597] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] max_logfile_count = 30 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.014757] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] max_logfile_size_mb = 200 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.014975] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] maximum_instance_delete_attempts = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.015173] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] metadata_listen = 0.0.0.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.015346] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] metadata_listen_port = 8775 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.015516] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] metadata_workers = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.015678] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] migrate_max_retries = -1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.015843] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] mkisofs_cmd = genisoimage {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.016092] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] my_block_storage_ip = 10.180.1.21 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.016237] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] my_ip = 10.180.1.21 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.016407] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] network_allocate_retries = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.016588] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.016758] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] osapi_compute_listen = 0.0.0.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.016924] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] osapi_compute_listen_port = 8774 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.017105] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] osapi_compute_unique_server_name_scope = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.017277] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] osapi_compute_workers = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.017441] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] password_length = 12 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.017608] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] periodic_enable = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.017764] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] periodic_fuzzy_delay = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.017928] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] pointer_model = usbtablet {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.018107] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] preallocate_images = none {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.018269] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] publish_errors = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.018399] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] pybasedir = /opt/stack/nova {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.018555] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ram_allocation_ratio = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.018713] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] rate_limit_burst = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.018880] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] rate_limit_except_level = CRITICAL {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.019048] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] rate_limit_interval = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.019213] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] reboot_timeout = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.019370] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] reclaim_instance_interval = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.019526] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] record = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.019694] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] reimage_timeout_per_gb = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.019859] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] report_interval = 120 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.020031] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] rescue_timeout = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.020196] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] reserved_host_cpus = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.020354] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] reserved_host_disk_mb = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.020513] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] reserved_host_memory_mb = 512 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.020671] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] reserved_huge_pages = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.020829] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] resize_confirm_window = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.020987] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] resize_fs_using_block_device = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.021158] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] resume_guests_state_on_host_boot = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.021380] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.021566] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] rpc_response_timeout = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.021824] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] run_external_periodic_tasks = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.021987] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] running_deleted_instance_action = reap {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.022167] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] running_deleted_instance_poll_interval = 1800 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.022333] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] running_deleted_instance_timeout = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.022494] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler_instance_sync_interval = 120 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.022662] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_down_time = 720 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.022832] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] servicegroup_driver = db {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.022991] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] shell_completion = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.023163] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] shelved_offload_time = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.023389] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] shelved_poll_interval = 3600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.023587] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] shutdown_timeout = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.023743] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] source_is_ipv6 = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.023968] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ssl_only = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.024233] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.024403] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] sync_power_state_interval = 600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.024565] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] sync_power_state_pool_size = 1000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.024735] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] syslog_log_facility = LOG_USER {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.024902] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] tempdir = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.025070] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] timeout_nbd = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.025246] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] transport_url = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.025412] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] update_resources_interval = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.025574] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] use_cow_images = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.025735] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] use_eventlog = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.026024] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] use_journal = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.026231] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] use_json = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.026397] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] use_rootwrap_daemon = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.026560] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] use_stderr = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.026723] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] use_syslog = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.026884] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vcpu_pin_set = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.027067] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plugging_is_fatal = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.027239] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plugging_timeout = 300 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.027407] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] virt_mkfs = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.027570] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] volume_usage_poll_interval = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.027731] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] watch_log_file = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.027898] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] web = /usr/share/spice-html5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 450.028138] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.028318] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.028488] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.028670] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_concurrency.disable_process_locking = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.029226] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.029427] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.029604] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.029781] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.029954] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.030141] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.030329] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.auth_strategy = keystone {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.030498] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.compute_link_prefix = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.030674] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.030853] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.dhcp_domain = novalocal {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.031035] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.enable_instance_password = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.031211] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.glance_link_prefix = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.031381] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.031555] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.031760] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.instance_list_per_project_cells = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.031966] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.list_records_by_skipping_down_cells = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.032158] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.local_metadata_per_cell = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.032334] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.max_limit = 1000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.032506] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.metadata_cache_expiration = 15 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.032685] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.neutron_default_tenant_id = default {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.032860] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.response_validation = warn {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.033052] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.use_neutron_default_nets = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.033232] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.033401] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.033572] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.033750] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.033967] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.vendordata_dynamic_targets = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.034195] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.vendordata_jsonfile_path = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.034392] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.034595] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.backend = dogpile.cache.memcached {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.034769] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.backend_argument = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.034945] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.config_prefix = cache.oslo {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.035133] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.dead_timeout = 60.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.035303] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.debug_cache_backend = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.035470] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.enable_retry_client = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.035633] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.enable_socket_keepalive = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.035808] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.enabled = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.035977] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.enforce_fips_mode = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.036157] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.expiration_time = 600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.036327] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.hashclient_retry_attempts = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.036499] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.hashclient_retry_delay = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.036665] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_dead_retry = 300 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.036828] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_password = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.037014] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.037178] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.037345] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_pool_maxsize = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.037511] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.037676] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_sasl_enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.037858] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.038040] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_socket_timeout = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.038210] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.memcache_username = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.038381] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.proxies = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.038549] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.redis_db = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.038713] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.redis_password = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.038886] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.redis_sentinel_service_name = mymaster {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.039079] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.039255] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.redis_server = localhost:6379 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.039431] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.redis_socket_timeout = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.039595] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.redis_username = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.039761] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.retry_attempts = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.039928] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.retry_delay = 0.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.040141] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.socket_keepalive_count = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.040327] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.socket_keepalive_idle = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.040496] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.socket_keepalive_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.040660] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.tls_allowed_ciphers = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.040822] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.tls_cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.040984] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.tls_certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.041163] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.tls_enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.041325] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cache.tls_keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.041499] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.041675] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.auth_type = password {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.041867] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.042063] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.catalog_info = volumev3::publicURL {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.042231] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.042399] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.042565] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.cross_az_attach = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.042729] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.debug = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.042891] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.endpoint_template = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.043068] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.http_retries = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.043237] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.043395] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.043568] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.os_region_name = RegionOne {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.043737] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.043934] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cinder.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.044120] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.044286] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.cpu_dedicated_set = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.044449] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.cpu_shared_set = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.044619] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.image_type_exclude_list = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.044787] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.044949] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.max_concurrent_disk_ops = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.045129] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.max_disk_devices_to_attach = -1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.045294] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.045465] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.045631] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.resource_provider_association_refresh = 300 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.045797] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.045960] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.shutdown_retry_interval = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.046154] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.046335] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] conductor.workers = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.046517] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] console.allowed_origins = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.046676] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] console.ssl_ciphers = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.046846] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] console.ssl_minimum_version = default {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.047028] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] consoleauth.enforce_session_timeout = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.047201] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] consoleauth.token_ttl = 600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.047371] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.047530] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.047694] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.047855] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.connect_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.048026] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.connect_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.048191] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.endpoint_override = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.048354] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.048514] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.048673] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.max_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.048832] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.min_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.048991] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.region_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.049160] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.retriable_status_codes = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.049319] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.service_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.049489] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.service_type = accelerator {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.049650] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.049808] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.status_code_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.049968] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.status_code_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.050141] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.050322] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.050485] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] cyborg.version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.050665] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.backend = sqlalchemy {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.050834] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.connection = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.051006] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.connection_debug = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.051184] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.connection_parameters = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.051351] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.connection_recycle_time = 3600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.051515] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.connection_trace = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.051681] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.db_inc_retry_interval = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.051908] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.db_max_retries = 20 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.052111] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.db_max_retry_interval = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.052288] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.db_retry_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.052455] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.max_overflow = 50 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.052619] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.max_pool_size = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.052782] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.max_retries = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.052954] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.053130] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.mysql_wsrep_sync_wait = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.053291] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.pool_timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.053455] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.retry_interval = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.053612] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.slave_connection = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.053773] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.sqlite_synchronous = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.053964] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] database.use_db_reconnect = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.054164] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.backend = sqlalchemy {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.054340] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.connection = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.054508] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.connection_debug = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.054681] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.connection_parameters = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.054850] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.connection_recycle_time = 3600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.055030] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.connection_trace = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.055198] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.db_inc_retry_interval = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.055366] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.db_max_retries = 20 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.055531] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.db_max_retry_interval = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.055695] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.db_retry_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.055861] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.max_overflow = 50 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.056035] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.max_pool_size = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.056203] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.max_retries = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.056376] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.056540] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.056701] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.pool_timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.056866] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.retry_interval = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.057043] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.slave_connection = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.057216] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] api_database.sqlite_synchronous = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.057396] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] devices.enabled_mdev_types = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.057574] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.057747] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ephemeral_storage_encryption.default_format = luks {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.057913] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ephemeral_storage_encryption.enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.058088] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.058265] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.api_servers = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.058430] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.058596] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.058760] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.058921] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.connect_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.059090] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.connect_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.059255] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.debug = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.059420] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.default_trusted_certificate_ids = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.059581] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.enable_certificate_validation = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.059739] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.enable_rbd_download = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.059897] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.endpoint_override = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.060069] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.060234] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.060392] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.max_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.060548] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.min_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.060709] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.num_retries = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.060876] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.rbd_ceph_conf = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.061045] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.rbd_connect_timeout = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.061214] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.rbd_pool = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.061377] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.rbd_user = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.061536] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.region_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.061700] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.retriable_status_codes = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.061895] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.service_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.062089] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.service_type = image {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.062256] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.062423] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.status_code_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.062581] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.status_code_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.062738] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.062917] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.063092] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.verify_glance_signatures = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.063253] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] glance.version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.063418] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] guestfs.debug = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.063583] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] mks.enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.063977] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.064194] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] image_cache.manager_interval = 2400 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.064369] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] image_cache.precache_concurrency = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.064539] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] image_cache.remove_unused_base_images = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.064709] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.064886] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.065067] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] image_cache.subdirectory_name = _base {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.065247] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.api_max_retries = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.065414] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.api_retry_interval = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.065574] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.065737] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.auth_type = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.065902] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.066072] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.066236] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.066397] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.conductor_group = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.066557] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.connect_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.066715] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.connect_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.066874] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.endpoint_override = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.067046] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.067209] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.067368] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.max_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.067526] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.min_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.067691] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.peer_list = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.067849] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.region_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.068015] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.retriable_status_codes = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.068186] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.serial_console_state_timeout = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.068345] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.service_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.068514] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.service_type = baremetal {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.068673] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.shard = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.068836] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.068995] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.status_code_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.069174] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.status_code_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.069332] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.069512] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.069677] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ironic.version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.069857] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.070042] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] key_manager.fixed_key = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.070231] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.070394] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.barbican_api_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.070554] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.barbican_endpoint = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.070722] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.barbican_endpoint_type = public {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.070880] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.barbican_region_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.071049] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.071212] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.071375] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.071536] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.071704] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.071904] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.number_of_retries = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.072093] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.retry_delay = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.072263] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.send_service_user_token = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.072427] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.072585] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.072746] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.verify_ssl = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.072902] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican.verify_ssl_path = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.073085] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.073255] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.auth_type = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.073415] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.073645] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.073852] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.074038] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.074208] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.074375] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.074535] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] barbican_service_user.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.074703] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.approle_role_id = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.074898] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.approle_secret_id = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.075118] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.kv_mountpoint = secret {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.075290] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.kv_path = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.075457] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.kv_version = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.075617] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.namespace = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.075775] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.root_token_id = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.075942] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.ssl_ca_crt_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.076156] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.timeout = 60.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.076328] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.use_ssl = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.076498] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.076671] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.076836] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.auth_type = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.076998] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.077170] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.077334] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.077492] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.connect_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.077652] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.connect_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.077812] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.endpoint_override = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.077973] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.078144] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.078303] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.max_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.078459] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.min_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.078616] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.region_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.078774] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.retriable_status_codes = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.078932] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.service_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.079115] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.service_type = identity {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.079280] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.079439] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.status_code_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.079600] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.status_code_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.079783] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.079987] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.080169] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] keystone.version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.080369] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.connection_uri = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.080535] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.cpu_mode = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.080702] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.cpu_model_extra_flags = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.080872] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.cpu_models = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.081057] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.cpu_power_governor_high = performance {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.081229] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.cpu_power_governor_low = powersave {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.081397] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.cpu_power_management = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.081567] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.081731] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.device_detach_attempts = 8 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.081920] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.device_detach_timeout = 20 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.082115] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.disk_cachemodes = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.082283] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.disk_prefix = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.082451] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.enabled_perf_events = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.082615] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.file_backed_memory = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.082782] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.gid_maps = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.082942] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.hw_disk_discard = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.083114] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.hw_machine_type = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.083288] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.images_rbd_ceph_conf = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.083455] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.083619] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.083883] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.images_rbd_glance_store_name = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.084097] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.images_rbd_pool = rbd {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.084277] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.images_type = default {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.084443] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.images_volume_group = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.084609] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.inject_key = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.084774] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.inject_partition = -2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.084937] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.inject_password = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.085114] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.iscsi_iface = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.085279] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.iser_use_multipath = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.085442] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_bandwidth = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.085607] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.085770] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_downtime = 500 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.085933] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.086106] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.086268] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_inbound_addr = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.086431] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.086595] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_permit_post_copy = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.086756] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_scheme = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.086931] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_timeout_action = abort {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.087104] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_tunnelled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.087266] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_uri = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.087428] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.live_migration_with_native_tls = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.087588] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.max_queues = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.087749] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.087995] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.088197] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.nfs_mount_options = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.088484] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.088662] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.088832] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.num_iser_scan_tries = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.088999] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.num_memory_encrypted_guests = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.089181] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.089347] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.num_pcie_ports = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.089514] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.num_volume_scan_tries = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.089682] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.pmem_namespaces = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.089878] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.quobyte_client_cfg = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.090172] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.090348] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rbd_connect_timeout = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.090514] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.090676] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.090836] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rbd_secret_uuid = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.090995] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rbd_user = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.091172] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.091341] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.remote_filesystem_transport = ssh {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.091502] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rescue_image_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.091659] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rescue_kernel_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.091829] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rescue_ramdisk_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.092031] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.092199] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.rx_queue_size = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.092368] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.smbfs_mount_options = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.092636] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.092809] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.snapshot_compression = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.092975] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.snapshot_image_format = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.093205] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.093372] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.sparse_logical_volumes = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.093535] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.swtpm_enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.093701] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.swtpm_group = tss {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.093891] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.swtpm_user = tss {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.094087] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.sysinfo_serial = unique {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.094253] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.tb_cache_size = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.094415] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.tx_queue_size = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.094584] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.uid_maps = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.094753] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.use_virtio_for_bridges = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.094925] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.virt_type = kvm {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.095108] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.volume_clear = zero {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.095277] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.volume_clear_size = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.095444] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.volume_use_multipath = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.095605] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.vzstorage_cache_path = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.095774] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.095942] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.vzstorage_mount_group = qemu {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.096125] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.vzstorage_mount_opts = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.096297] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.096566] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.096743] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.vzstorage_mount_user = stack {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.096911] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.097103] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.097283] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.auth_type = password {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.097447] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.097609] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.097774] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.097936] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.connect_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.098109] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.connect_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.098284] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.default_floating_pool = public {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.098442] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.endpoint_override = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.098602] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.extension_sync_interval = 600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.098766] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.http_retries = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.098927] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.099100] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.099263] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.max_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.099432] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.099593] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.min_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.099760] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.ovs_bridge = br-int {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.099927] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.physnets = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.100108] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.region_name = RegionOne {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.100270] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.retriable_status_codes = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.100437] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.service_metadata_proxy = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.100596] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.service_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.100763] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.service_type = network {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.100922] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.101095] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.status_code_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.101258] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.status_code_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.101419] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.101600] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.101779] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] neutron.version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.101965] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] notifications.bdms_in_notifications = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.102159] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] notifications.default_level = INFO {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.102338] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] notifications.notification_format = unversioned {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.102502] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] notifications.notify_on_state_change = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.102678] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.102856] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] pci.alias = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.103041] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] pci.device_spec = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.103214] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] pci.report_in_placement = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.103390] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.103565] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.auth_type = password {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.103737] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.103921] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.104104] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.104274] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.104434] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.connect_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.104595] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.connect_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.104756] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.default_domain_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.104929] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.default_domain_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.105116] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.domain_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.105282] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.domain_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.105443] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.endpoint_override = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.105605] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.105764] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.105923] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.max_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.106101] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.min_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.106274] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.password = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.106436] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.project_domain_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.106603] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.project_domain_name = Default {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.106777] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.project_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.106949] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.project_name = service {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.107133] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.region_name = RegionOne {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.107298] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.retriable_status_codes = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.107460] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.service_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.107630] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.service_type = placement {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.107795] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.107956] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.status_code_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.108133] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.status_code_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.108299] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.system_scope = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.108461] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.108621] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.trust_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.108780] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.user_domain_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.108950] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.user_domain_name = Default {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.109125] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.user_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.109300] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.username = nova {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.109485] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.109649] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] placement.version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.109828] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.cores = 20 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.109995] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.count_usage_from_placement = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.110182] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.110359] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.injected_file_content_bytes = 10240 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.110528] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.injected_file_path_length = 255 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.110697] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.injected_files = 5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.110864] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.instances = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.111043] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.key_pairs = 100 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.111216] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.metadata_items = 128 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.111384] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.ram = 51200 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.111547] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.recheck_quota = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.111729] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.server_group_members = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.111914] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] quota.server_groups = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.112109] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.112278] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.112442] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.image_metadata_prefilter = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.112607] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.112772] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.max_attempts = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.112938] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.max_placement_results = 1000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.113143] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.113314] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.query_placement_for_image_type_support = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.113478] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.113649] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] scheduler.workers = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.113849] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.114038] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.114225] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.114399] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.114568] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.114735] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.114902] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.115106] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.115281] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.host_subset_size = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.115446] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.115606] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.115773] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.115935] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.isolated_hosts = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.116112] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.isolated_images = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.116276] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.116435] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.116599] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.116761] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.pci_in_placement = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.116926] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.117101] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.117268] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.117428] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.117590] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.117754] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.117916] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.track_instance_changes = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.118105] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.118282] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] metrics.required = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.118449] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] metrics.weight_multiplier = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.118612] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.118774] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] metrics.weight_setting = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.119091] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.119273] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] serial_console.enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.119448] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] serial_console.port_range = 10000:20000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.119619] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.119788] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.119958] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] serial_console.serialproxy_port = 6083 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.120145] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.120323] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.auth_type = password {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.120486] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.120646] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.120809] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.120973] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.121146] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.121317] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.send_service_user_token = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.121481] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.121654] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] service_user.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.121858] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.agent_enabled = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.122037] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.122357] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.122552] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.122723] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.html5proxy_port = 6082 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.122887] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.image_compression = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.123087] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.jpeg_compression = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.123299] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.playback_compression = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.123474] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.require_secure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.123648] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.server_listen = 127.0.0.1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.123835] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.124015] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.streaming_mode = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.124189] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] spice.zlib_compression = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.124359] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] upgrade_levels.baseapi = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.124529] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] upgrade_levels.compute = auto {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.124688] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] upgrade_levels.conductor = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.124845] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] upgrade_levels.scheduler = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.125022] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.125191] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.auth_type = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.125351] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.125509] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.125669] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.125829] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.125989] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.126165] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.126325] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vendordata_dynamic_auth.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.126500] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.api_retry_count = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.126664] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.ca_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.126835] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.cache_prefix = devstack-image-cache {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.127013] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.cluster_name = testcl1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.127184] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.connection_pool_size = 10 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.127344] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.console_delay_seconds = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.127511] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.datastore_regex = ^datastore.* {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.127712] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.127885] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.host_password = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.128062] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.host_port = 443 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.128236] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.host_username = administrator@vsphere.local {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.128404] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.insecure = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.128567] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.integration_bridge = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.128731] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.maximum_objects = 100 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.128890] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.pbm_default_policy = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.129065] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.pbm_enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.129227] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.pbm_wsdl_location = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.129396] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.129556] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.serial_port_proxy_uri = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.129713] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.serial_port_service_uri = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.129921] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.task_poll_interval = 0.5 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.130157] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.use_linked_clone = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.130342] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.vnc_keymap = en-us {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.130513] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.vnc_port = 5900 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.130680] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vmware.vnc_port_total = 10000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.130871] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.auth_schemes = ['none'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.131058] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.131350] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.131537] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.131727] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.novncproxy_port = 6080 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.131933] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.server_listen = 127.0.0.1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.132121] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.132289] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.vencrypt_ca_certs = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.132452] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.vencrypt_client_cert = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.132612] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vnc.vencrypt_client_key = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.132792] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.132958] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.disable_deep_image_inspection = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.133203] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.133385] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.133552] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.133719] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.disable_rootwrap = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.133911] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.enable_numa_live_migration = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.134099] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.134273] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.134438] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.134604] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.libvirt_disable_apic = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.134767] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.134933] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.135110] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.135276] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.135440] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.135602] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.135764] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.135925] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.136112] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.136342] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.136536] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.136711] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.client_socket_timeout = 900 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.136879] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.default_pool_size = 1000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.137061] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.keep_alive = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.137234] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.max_header_line = 16384 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.137398] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.secure_proxy_ssl_header = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.137559] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.ssl_ca_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.137719] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.ssl_cert_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.137880] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.ssl_key_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.138056] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.tcp_keepidle = 600 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.138244] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.138414] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] zvm.ca_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.138578] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] zvm.cloud_connector_url = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.138857] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.139048] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] zvm.reachable_timeout = 300 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.139269] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.enforce_new_defaults = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.139676] env[61768]: WARNING oslo_config.cfg [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 450.139866] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.enforce_scope = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.140059] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.policy_default_rule = default {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.140249] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.140426] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.policy_file = policy.yaml {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.140603] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.140766] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.140926] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.141098] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.141264] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.141436] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.141614] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.141822] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.connection_string = messaging:// {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.141998] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.enabled = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.142184] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.es_doc_type = notification {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.142352] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.es_scroll_size = 10000 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.142522] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.es_scroll_time = 2m {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.142684] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.filter_error_trace = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.142852] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.hmac_keys = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.143028] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.sentinel_service_name = mymaster {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.143201] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.socket_timeout = 0.1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.143363] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.trace_requests = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.143523] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler.trace_sqlalchemy = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.143704] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler_jaeger.process_tags = {} {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.143890] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler_jaeger.service_name_prefix = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.144074] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] profiler_otlp.service_name_prefix = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.144246] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] remote_debug.host = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.144411] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] remote_debug.port = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.144590] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.144753] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.144937] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.145180] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.145364] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.145530] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.145696] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.145861] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.146039] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.146214] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.146375] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.146544] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.146713] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.146884] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.147065] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.147236] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.147401] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.147574] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.147739] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.147904] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.148088] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.148262] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.148427] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.148596] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.148761] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.148927] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.149104] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.149272] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.149444] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.149611] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.ssl = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.149786] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.149958] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.150140] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.150313] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.150485] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.ssl_version = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.150650] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.150835] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.151013] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_notifications.retry = -1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.151205] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.151385] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_messaging_notifications.transport_url = **** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.151559] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.auth_section = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.151747] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.auth_type = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.151924] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.cafile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.152100] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.certfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.152267] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.collect_timing = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.152426] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.connect_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.152587] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.connect_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.152745] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.endpoint_id = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.152903] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.endpoint_override = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.153075] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.insecure = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.153236] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.keyfile = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.153394] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.max_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.153550] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.min_version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.153707] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.region_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.153890] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.retriable_status_codes = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.154073] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.service_name = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.154237] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.service_type = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.154400] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.split_loggers = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.154557] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.status_code_retries = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.154715] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.status_code_retry_delay = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.154878] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.timeout = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.155072] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.valid_interfaces = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.155236] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_limit.version = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.155404] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_reports.file_event_handler = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.155570] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.155730] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] oslo_reports.log_dir = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.155902] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.156074] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.156238] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.156404] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.156572] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.156733] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.156905] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.157110] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_ovs_privileged.group = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.157279] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.157449] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.157615] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.157776] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] vif_plug_ovs_privileged.user = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.157951] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.flat_interface = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.158148] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.158327] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.158500] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.158670] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.158837] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.159014] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.159190] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.159370] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.159544] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_ovs.isolate_vif = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.159713] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.159878] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.160057] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.160236] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_ovs.ovsdb_interface = native {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.160399] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] os_vif_ovs.per_port_bridge = False {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.160569] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] privsep_osbrick.capabilities = [21] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.160728] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] privsep_osbrick.group = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.160886] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] privsep_osbrick.helper_command = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.161060] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.161226] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.161383] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] privsep_osbrick.user = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.161555] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.161733] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] nova_sys_admin.group = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.161910] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] nova_sys_admin.helper_command = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.162094] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.162268] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.162445] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] nova_sys_admin.user = None {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 450.162576] env[61768]: DEBUG oslo_service.service [None req-85ec4605-19de-421d-8c72-d1496be75b97 None None] ******************************************************************************** {{(pid=61768) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 450.163057] env[61768]: INFO nova.service [-] Starting compute node (version 30.1.0) [ 450.666571] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Getting list of instances from cluster (obj){ [ 450.666571] env[61768]: value = "domain-c8" [ 450.666571] env[61768]: _type = "ClusterComputeResource" [ 450.666571] env[61768]: } {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 450.667709] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5957ba-3c24-4176-bf45-1849546eba99 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 450.676736] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Got total of 0 instances {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 450.677261] env[61768]: WARNING nova.virt.vmwareapi.driver [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 450.677712] env[61768]: INFO nova.virt.node [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Generated node identity f9f7149c-7c4e-44e7-8000-07a202159d2f [ 450.677940] env[61768]: INFO nova.virt.node [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Wrote node identity f9f7149c-7c4e-44e7-8000-07a202159d2f to /opt/stack/data/n-cpu-1/compute_id [ 451.180645] env[61768]: WARNING nova.compute.manager [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Compute nodes ['f9f7149c-7c4e-44e7-8000-07a202159d2f'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 452.186649] env[61768]: INFO nova.compute.manager [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 453.192737] env[61768]: WARNING nova.compute.manager [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 453.193144] env[61768]: DEBUG oslo_concurrency.lockutils [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 453.193266] env[61768]: DEBUG oslo_concurrency.lockutils [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 453.193424] env[61768]: DEBUG oslo_concurrency.lockutils [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 453.193587] env[61768]: DEBUG nova.compute.resource_tracker [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 453.194556] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4277c2c2-c6e1-4bd3-a678-2514a8155f42 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.203011] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-849a0daf-fd18-4ccd-858e-2503e65ba533 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.216305] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f98bd8-9bf1-4e0c-8329-2665d7540d28 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.222591] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbf5cb4b-e197-4848-81dc-18fa038d4795 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 453.251766] env[61768]: DEBUG nova.compute.resource_tracker [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181504MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 453.251928] env[61768]: DEBUG oslo_concurrency.lockutils [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 453.252111] env[61768]: DEBUG oslo_concurrency.lockutils [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 453.754822] env[61768]: WARNING nova.compute.resource_tracker [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] No compute node record for cpu-1:f9f7149c-7c4e-44e7-8000-07a202159d2f: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host f9f7149c-7c4e-44e7-8000-07a202159d2f could not be found. [ 454.258519] env[61768]: INFO nova.compute.resource_tracker [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: f9f7149c-7c4e-44e7-8000-07a202159d2f [ 455.766444] env[61768]: DEBUG nova.compute.resource_tracker [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 455.766844] env[61768]: DEBUG nova.compute.resource_tracker [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 455.917613] env[61768]: INFO nova.scheduler.client.report [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] [req-af79e1d3-85d1-48f7-8766-238223d87583] Created resource provider record via placement API for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 455.929822] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37734045-0632-4e72-853e-9222072e2c7e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.937395] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8492132-4a48-4ff3-8e0e-1e215b09be63 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.966681] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de60aa7-0100-4708-8e62-cc6c2672c703 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.973391] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f842ee-f47f-4039-802b-2fc4bbcca9e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 455.985809] env[61768]: DEBUG nova.compute.provider_tree [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 456.520252] env[61768]: DEBUG nova.scheduler.client.report [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Updated inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 456.520489] env[61768]: DEBUG nova.compute.provider_tree [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 0 to 1 during operation: update_inventory {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 456.520634] env[61768]: DEBUG nova.compute.provider_tree [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 456.570869] env[61768]: DEBUG nova.compute.provider_tree [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 1 to 2 during operation: update_traits {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 457.075189] env[61768]: DEBUG nova.compute.resource_tracker [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 457.075558] env[61768]: DEBUG oslo_concurrency.lockutils [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.823s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 457.075558] env[61768]: DEBUG nova.service [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Creating RPC server for service compute {{(pid=61768) start /opt/stack/nova/nova/service.py:186}} [ 457.089672] env[61768]: DEBUG nova.service [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] Join ServiceGroup membership for this service compute {{(pid=61768) start /opt/stack/nova/nova/service.py:203}} [ 457.089866] env[61768]: DEBUG nova.servicegroup.drivers.db [None req-634b9ef6-b497-4990-aa84-136187d10bc6 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61768) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 480.096470] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_power_states {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 480.599266] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Getting list of instances from cluster (obj){ [ 480.599266] env[61768]: value = "domain-c8" [ 480.599266] env[61768]: _type = "ClusterComputeResource" [ 480.599266] env[61768]: } {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 480.600461] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7121e7ac-e30b-4198-8393-251183dcd3c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.610344] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Got total of 0 instances {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 480.610596] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 480.610914] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Getting list of instances from cluster (obj){ [ 480.610914] env[61768]: value = "domain-c8" [ 480.610914] env[61768]: _type = "ClusterComputeResource" [ 480.610914] env[61768]: } {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 480.611815] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a2f47b-40b0-4c48-8b43-0914b8e23739 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.624793] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Got total of 0 instances {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 492.017077] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Acquiring lock "1efe0c6f-38e1-4131-9adc-995cbf33fa0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 492.017380] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Lock "1efe0c6f-38e1-4131-9adc-995cbf33fa0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 492.521426] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 493.073075] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 493.073385] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 493.074977] env[61768]: INFO nova.compute.claims [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 494.124434] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57484fba-af89-439f-933c-cc393e0048eb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.133588] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f892c007-6d0b-43fe-911e-ad3336595a9f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.172016] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afc1482d-24a1-4ad3-ac50-5d4693fb99b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.183367] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5383ff5-31a1-4d5f-9188-7fadcfc66695 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 494.196874] env[61768]: DEBUG nova.compute.provider_tree [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 494.621393] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Acquiring lock "0b1de80e-6342-4443-a700-3d50f28270b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.624733] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Lock "0b1de80e-6342-4443-a700-3d50f28270b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 494.703217] env[61768]: DEBUG nova.scheduler.client.report [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 494.851570] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "df682d97-53c1-4a18-8800-b23f3fc66018" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 494.852016] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "df682d97-53c1-4a18-8800-b23f3fc66018" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.129246] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 495.173535] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "ca465e62-c190-4d67-9dec-87ea1aa8f194" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.173686] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "ca465e62-c190-4d67-9dec-87ea1aa8f194" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.206807] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.133s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 495.207548] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 495.357243] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 495.658484] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.658727] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.660799] env[61768]: INFO nova.compute.claims [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 495.676500] env[61768]: DEBUG nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 495.718048] env[61768]: DEBUG nova.compute.utils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 495.721741] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 495.723147] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 495.875426] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "438d040b-c86f-47ab-8d22-e0e86c41ac43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 495.875691] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "438d040b-c86f-47ab-8d22-e0e86c41ac43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 495.907891] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.153967] env[61768]: DEBUG nova.policy [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5fb38942aaf64900b381ada6c8bd2122', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'da629d6efbf04ba38429bd5a6c031e17', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 496.220374] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 496.230305] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 496.380709] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 496.804023] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118bffa6-348a-4b19-9a37-fc87b6a6fe84 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.816157] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd71d297-fd1d-440d-8bd7-8cb429b9b36b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.855386] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06f78225-3591-4716-933c-2357e45654c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.866932] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc69558-96f0-48ec-9b78-679ebca9bb80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 496.881177] env[61768]: DEBUG nova.compute.provider_tree [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 496.923118] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 497.240674] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 497.271240] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 497.271529] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 497.271701] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 497.271896] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 497.272118] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 497.272312] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 497.272521] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 497.272748] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 497.274372] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 497.274372] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 497.274372] env[61768]: DEBUG nova.virt.hardware [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 497.274646] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f9720a2-4662-4811-baa1-45d9c90396a2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.283380] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d8eaa1-dad3-4020-9f60-4d2462e95040 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.302559] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58b25349-1d70-41b2-9589-ca4a9cbd08e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.385562] env[61768]: DEBUG nova.scheduler.client.report [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 497.542411] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Successfully created port: 6f476265-1e10-4150-83d5-d936bb9146dc {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 497.893354] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.234s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 497.894472] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 497.898047] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.990s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 497.903025] env[61768]: INFO nova.compute.claims [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 498.408358] env[61768]: DEBUG nova.compute.utils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 498.413152] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 498.413781] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 498.481702] env[61768]: DEBUG nova.policy [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9cd45a8cebcf43298c40353368857f67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c5d087105a3f4fb2a83843d19dd417d9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 498.917956] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 499.035281] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c378d80-12a9-4caf-a9a2-8c7fa9507ec3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.041456] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8fc3251-7a9c-4bf0-955a-9a998a6db9e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.076067] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5c8321-8d51-48b6-881b-83ab0e351d06 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.079410] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Successfully created port: 8fe593a8-1b08-4764-8ba1-e45d2658db24 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 499.086684] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad6c9d8-baf1-4695-8c5f-19f705478600 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.104616] env[61768]: DEBUG nova.compute.provider_tree [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 499.611098] env[61768]: DEBUG nova.scheduler.client.report [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 499.931753] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 499.968223] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 499.968223] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 499.968223] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 499.968430] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 499.968430] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 499.968430] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 499.968569] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 499.968716] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 499.968876] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 499.969101] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 499.969294] env[61768]: DEBUG nova.virt.hardware [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 499.970200] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0875c09-da72-476f-bed3-95e03ce5e948 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.978793] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c2529d4-a7fc-48b1-b9f9-0490543ec4a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 500.119642] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.221s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 500.119642] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 500.124823] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.905s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 500.126744] env[61768]: INFO nova.compute.claims [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 500.631365] env[61768]: DEBUG nova.compute.utils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 500.636443] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 500.636747] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 500.878759] env[61768]: DEBUG nova.policy [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b644f2c92abd4559a88d99199314268f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae392a1509054e6ab7e8698b20db16ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 501.107360] env[61768]: ERROR nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. [ 501.107360] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 501.107360] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.107360] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 501.107360] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 501.107360] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 501.107360] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 501.107360] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 501.107360] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.107360] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 501.107360] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.107360] env[61768]: ERROR nova.compute.manager raise self.value [ 501.107360] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 501.107360] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 501.107360] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.107360] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 501.107968] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.107968] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 501.107968] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. [ 501.107968] env[61768]: ERROR nova.compute.manager [ 501.107968] env[61768]: Traceback (most recent call last): [ 501.107968] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 501.107968] env[61768]: listener.cb(fileno) [ 501.107968] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.107968] env[61768]: result = function(*args, **kwargs) [ 501.107968] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 501.107968] env[61768]: return func(*args, **kwargs) [ 501.107968] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.107968] env[61768]: raise e [ 501.107968] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.107968] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 501.107968] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 501.107968] env[61768]: created_port_ids = self._update_ports_for_instance( [ 501.107968] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 501.107968] env[61768]: with excutils.save_and_reraise_exception(): [ 501.107968] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.107968] env[61768]: self.force_reraise() [ 501.107968] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.107968] env[61768]: raise self.value [ 501.107968] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 501.107968] env[61768]: updated_port = self._update_port( [ 501.107968] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.107968] env[61768]: _ensure_no_port_binding_failure(port) [ 501.107968] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.107968] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 501.108828] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. [ 501.108828] env[61768]: Removing descriptor: 14 [ 501.109276] env[61768]: ERROR nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Traceback (most recent call last): [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] yield resources [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self.driver.spawn(context, instance, image_meta, [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] vm_ref = self.build_virtual_machine(instance, [ 501.109276] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] for vif in network_info: [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return self._sync_wrapper(fn, *args, **kwargs) [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self.wait() [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self[:] = self._gt.wait() [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return self._exit_event.wait() [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 501.109606] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] result = hub.switch() [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return self.greenlet.switch() [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] result = function(*args, **kwargs) [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return func(*args, **kwargs) [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] raise e [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] nwinfo = self.network_api.allocate_for_instance( [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] created_port_ids = self._update_ports_for_instance( [ 501.110443] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] with excutils.save_and_reraise_exception(): [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self.force_reraise() [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] raise self.value [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] updated_port = self._update_port( [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] _ensure_no_port_binding_failure(port) [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] raise exception.PortBindingFailed(port_id=port['id']) [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] nova.exception.PortBindingFailed: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. [ 501.111064] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] [ 501.112372] env[61768]: INFO nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Terminating instance [ 501.112372] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Acquiring lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 501.112372] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Acquired lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 501.112372] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 501.136893] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 501.240533] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7750fb8c-b8d6-44d6-af53-ec18e911f81e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.251052] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0d6266-95c8-4e61-805d-3e0f8b8c4f8e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.290278] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ae4135-86f5-4158-a71b-cbda1b3b7c1d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.299065] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ac3142-2213-4ca2-bc17-278a71c5b560 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 501.317738] env[61768]: DEBUG nova.compute.provider_tree [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 501.525672] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Acquiring lock "0188dbb6-ae91-4601-83a3-3e3bf55fc753" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 501.527100] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Lock "0188dbb6-ae91-4601-83a3-3e3bf55fc753" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 501.721237] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 501.728020] env[61768]: ERROR nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. [ 501.728020] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 501.728020] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.728020] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 501.728020] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 501.728020] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 501.728020] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 501.728020] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 501.728020] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.728020] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 501.728020] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.728020] env[61768]: ERROR nova.compute.manager raise self.value [ 501.728020] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 501.728020] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 501.728020] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.728020] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 501.728516] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.728516] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 501.728516] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. [ 501.728516] env[61768]: ERROR nova.compute.manager [ 501.728516] env[61768]: Traceback (most recent call last): [ 501.728516] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 501.728516] env[61768]: listener.cb(fileno) [ 501.728516] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.728516] env[61768]: result = function(*args, **kwargs) [ 501.728516] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 501.728516] env[61768]: return func(*args, **kwargs) [ 501.728516] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.728516] env[61768]: raise e [ 501.728516] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.728516] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 501.728516] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 501.728516] env[61768]: created_port_ids = self._update_ports_for_instance( [ 501.728516] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 501.728516] env[61768]: with excutils.save_and_reraise_exception(): [ 501.728516] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.728516] env[61768]: self.force_reraise() [ 501.728516] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.728516] env[61768]: raise self.value [ 501.728516] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 501.728516] env[61768]: updated_port = self._update_port( [ 501.728516] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.728516] env[61768]: _ensure_no_port_binding_failure(port) [ 501.728516] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.728516] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 501.729207] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. [ 501.729207] env[61768]: Removing descriptor: 16 [ 501.729207] env[61768]: ERROR nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Traceback (most recent call last): [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] yield resources [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self.driver.spawn(context, instance, image_meta, [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 501.729207] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] vm_ref = self.build_virtual_machine(instance, [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] for vif in network_info: [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return self._sync_wrapper(fn, *args, **kwargs) [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self.wait() [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self[:] = self._gt.wait() [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return self._exit_event.wait() [ 501.729515] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] result = hub.switch() [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return self.greenlet.switch() [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] result = function(*args, **kwargs) [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return func(*args, **kwargs) [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] raise e [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] nwinfo = self.network_api.allocate_for_instance( [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 501.729870] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] created_port_ids = self._update_ports_for_instance( [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] with excutils.save_and_reraise_exception(): [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self.force_reraise() [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] raise self.value [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] updated_port = self._update_port( [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] _ensure_no_port_binding_failure(port) [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 501.730286] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] raise exception.PortBindingFailed(port_id=port['id']) [ 501.730579] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] nova.exception.PortBindingFailed: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. [ 501.730579] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] [ 501.730579] env[61768]: INFO nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Terminating instance [ 501.731916] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Acquiring lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 501.732113] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Acquired lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 501.732287] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 501.823553] env[61768]: DEBUG nova.scheduler.client.report [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 502.034043] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 502.090251] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 502.149449] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 502.183724] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 502.183974] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 502.184145] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 502.184331] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 502.184476] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 502.184622] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 502.184841] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 502.184992] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 502.186589] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 502.186589] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 502.186589] env[61768]: DEBUG nova.virt.hardware [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 502.186916] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7edb6cd3-4eda-4d31-a79e-cf5eb81a3939 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.197027] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb83aa2d-6770-4212-8059-51ef0f63f3d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.220621] env[61768]: DEBUG nova.compute.manager [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Received event network-changed-6f476265-1e10-4150-83d5-d936bb9146dc {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 502.220957] env[61768]: DEBUG nova.compute.manager [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Refreshing instance network info cache due to event network-changed-6f476265-1e10-4150-83d5-d936bb9146dc. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 502.221196] env[61768]: DEBUG oslo_concurrency.lockutils [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] Acquiring lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.222011] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Successfully created port: c3e98df6-99ea-47a5-81e3-0fb22025e494 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 502.266371] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 502.330566] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.206s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 502.331086] env[61768]: DEBUG nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 502.337037] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.412s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 502.337037] env[61768]: INFO nova.compute.claims [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 502.400924] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 502.500127] env[61768]: DEBUG nova.compute.manager [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Received event network-changed-8fe593a8-1b08-4764-8ba1-e45d2658db24 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 502.500127] env[61768]: DEBUG nova.compute.manager [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Refreshing instance network info cache due to event network-changed-8fe593a8-1b08-4764-8ba1-e45d2658db24. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 502.500127] env[61768]: DEBUG oslo_concurrency.lockutils [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] Acquiring lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 502.571825] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 502.601196] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Releasing lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 502.601196] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 502.601196] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 502.601196] env[61768]: DEBUG oslo_concurrency.lockutils [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] Acquired lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 502.601395] env[61768]: DEBUG nova.network.neutron [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Refreshing network info cache for port 6f476265-1e10-4150-83d5-d936bb9146dc {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 502.602433] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a8a23c25-86ec-4d4f-82dd-cc1810e37e50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.619044] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113222b9-fc02-4eb0-8665-3fc87e7d7eba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.649667] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1efe0c6f-38e1-4131-9adc-995cbf33fa0c could not be found. [ 502.650090] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 502.650854] env[61768]: INFO nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Took 0.05 seconds to destroy the instance on the hypervisor. [ 502.650854] env[61768]: DEBUG oslo.service.loopingcall [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 502.650854] env[61768]: DEBUG nova.compute.manager [-] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 502.651246] env[61768]: DEBUG nova.network.neutron [-] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 502.703095] env[61768]: DEBUG nova.network.neutron [-] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 502.840887] env[61768]: DEBUG nova.compute.utils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 502.845884] env[61768]: DEBUG nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Not allocating networking since 'none' was specified. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 502.905220] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Releasing lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 502.905220] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 502.905220] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 502.905220] env[61768]: DEBUG oslo_concurrency.lockutils [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] Acquired lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 502.905220] env[61768]: DEBUG nova.network.neutron [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Refreshing network info cache for port 8fe593a8-1b08-4764-8ba1-e45d2658db24 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 502.906302] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4d2623d-9c1e-4922-97f5-cdb1ecf5c33a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.915719] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e06ed4d-0b7f-485b-b209-277a4cd44bc4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 502.940008] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0b1de80e-6342-4443-a700-3d50f28270b6 could not be found. [ 502.940343] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 502.940530] env[61768]: INFO nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 502.940914] env[61768]: DEBUG oslo.service.loopingcall [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 502.941131] env[61768]: DEBUG nova.compute.manager [-] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 502.941220] env[61768]: DEBUG nova.network.neutron [-] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 502.992385] env[61768]: DEBUG nova.network.neutron [-] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 503.147764] env[61768]: DEBUG nova.network.neutron [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 503.205702] env[61768]: DEBUG nova.network.neutron [-] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.320275] env[61768]: DEBUG nova.network.neutron [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.347207] env[61768]: DEBUG nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 503.456830] env[61768]: DEBUG nova.network.neutron [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 503.477521] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-649d2383-e887-4d90-aac6-4c20b68d0180 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.485572] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0540382-5e0f-460c-b72e-b4ef4e7097b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.520024] env[61768]: DEBUG nova.network.neutron [-] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.522625] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86f0663-4570-4100-b2a9-097061c0a56f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.533695] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ccac13c-cfb9-4813-8ae7-d9e38f9887c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 503.552828] env[61768]: DEBUG nova.compute.provider_tree [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 503.687887] env[61768]: DEBUG nova.network.neutron [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 503.710755] env[61768]: INFO nova.compute.manager [-] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Took 1.06 seconds to deallocate network for instance. [ 503.713394] env[61768]: DEBUG nova.compute.claims [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 503.713457] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 503.823717] env[61768]: DEBUG oslo_concurrency.lockutils [req-bd4359a1-5520-40b7-9e16-4089b020e341 req-26130382-5552-4b2f-a617-94bd5c4be261 service nova] Releasing lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 504.026858] env[61768]: INFO nova.compute.manager [-] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Took 1.09 seconds to deallocate network for instance. [ 504.030715] env[61768]: DEBUG nova.compute.claims [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 504.030908] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.056365] env[61768]: DEBUG nova.scheduler.client.report [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 504.192486] env[61768]: DEBUG oslo_concurrency.lockutils [req-15a8cf28-7fed-45ca-911b-20101fdb64b0 req-4baa1082-1a97-4a14-9e4a-d057e21c7ac8 service nova] Releasing lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 504.280011] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "c3b90a4d-e885-453c-9f8e-799adb2400a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.280267] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "c3b90a4d-e885-453c-9f8e-799adb2400a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.362555] env[61768]: DEBUG nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 504.400871] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 504.400871] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 504.400871] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 504.401154] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 504.401154] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 504.401154] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 504.401154] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 504.401432] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 504.401482] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 504.401635] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 504.401800] env[61768]: DEBUG nova.virt.hardware [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 504.402975] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964f5c6c-8a26-4f13-ab4a-f5e38b0accb6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.415692] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddb2602-a0c9-4be7-97b2-e9704b96b439 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.438632] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 504.452575] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 504.452728] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7cd380ea-d318-4e15-a740-889570b4db9b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.470540] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Created folder: OpenStack in parent group-v4. [ 504.470940] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Creating folder: Project (30bfaf5d9c3042a4a9fc3b2124a1bc83). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 504.470940] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3bf8cd26-45cc-47da-88bb-8bb2ad1ff788 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.482762] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Created folder: Project (30bfaf5d9c3042a4a9fc3b2124a1bc83) in parent group-v265360. [ 504.483397] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Creating folder: Instances. Parent ref: group-v265361. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 504.483397] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcddbdb2-32d9-4bf2-9b88-42f8dd7ca6c3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.499161] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Created folder: Instances in parent group-v265361. [ 504.499161] env[61768]: DEBUG oslo.service.loopingcall [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 504.499161] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 504.499161] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7c325894-82b3-4919-aa81-21df226e9acf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 504.524497] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 504.524497] env[61768]: value = "task-1228511" [ 504.524497] env[61768]: _type = "Task" [ 504.524497] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 504.535121] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228511, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 504.560935] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.226s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 504.561553] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 504.570119] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.995s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.571683] env[61768]: INFO nova.compute.claims [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 504.693548] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 504.693548] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 504.784035] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 505.035074] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228511, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.085515] env[61768]: DEBUG nova.compute.utils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 505.092066] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 505.095212] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 505.199261] env[61768]: DEBUG nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 505.246363] env[61768]: DEBUG nova.policy [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f538a940471d4bc3941415c41d4ef9d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f989ffd7857b42599b4e3585d7f39506', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 505.310751] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.377144] env[61768]: DEBUG nova.compute.manager [req-b3ecf9f8-0e1b-433a-8a98-3662d308b6c1 req-70a99bb9-53e4-406c-8ba0-5ba559fb9b69 service nova] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Received event network-vif-deleted-6f476265-1e10-4150-83d5-d936bb9146dc {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 505.536218] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228511, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 505.593308] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 505.629375] env[61768]: DEBUG nova.compute.manager [req-d0d36424-b48f-48bc-86d5-b281fb0c60e7 req-439867b8-8f82-4baa-9d4d-8af24b00b4fa service nova] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Received event network-vif-deleted-8fe593a8-1b08-4764-8ba1-e45d2658db24 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 505.629787] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 505.636225] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 505.636463] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 505.636599] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Rebuilding the list of instances to heal {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 505.726696] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 505.769910] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4116319-0014-4b4d-abe6-48a60940c456 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.775694] env[61768]: ERROR nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. [ 505.775694] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 505.775694] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 505.775694] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 505.775694] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 505.775694] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 505.775694] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 505.775694] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 505.775694] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.775694] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 505.775694] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.775694] env[61768]: ERROR nova.compute.manager raise self.value [ 505.775694] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 505.775694] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 505.775694] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.775694] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 505.776275] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.776275] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 505.776275] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. [ 505.776275] env[61768]: ERROR nova.compute.manager [ 505.776275] env[61768]: Traceback (most recent call last): [ 505.776275] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 505.776275] env[61768]: listener.cb(fileno) [ 505.776275] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 505.776275] env[61768]: result = function(*args, **kwargs) [ 505.776275] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 505.776275] env[61768]: return func(*args, **kwargs) [ 505.776275] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 505.776275] env[61768]: raise e [ 505.776275] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 505.776275] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 505.776275] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 505.776275] env[61768]: created_port_ids = self._update_ports_for_instance( [ 505.776275] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 505.776275] env[61768]: with excutils.save_and_reraise_exception(): [ 505.776275] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.776275] env[61768]: self.force_reraise() [ 505.776275] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.776275] env[61768]: raise self.value [ 505.776275] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 505.776275] env[61768]: updated_port = self._update_port( [ 505.776275] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.776275] env[61768]: _ensure_no_port_binding_failure(port) [ 505.776275] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.776275] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 505.777110] env[61768]: nova.exception.PortBindingFailed: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. [ 505.777110] env[61768]: Removing descriptor: 17 [ 505.777110] env[61768]: ERROR nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Traceback (most recent call last): [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] yield resources [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self.driver.spawn(context, instance, image_meta, [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self._vmops.spawn(context, instance, image_meta, injected_files, [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 505.777110] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] vm_ref = self.build_virtual_machine(instance, [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] vif_infos = vmwarevif.get_vif_info(self._session, [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] for vif in network_info: [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return self._sync_wrapper(fn, *args, **kwargs) [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self.wait() [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self[:] = self._gt.wait() [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return self._exit_event.wait() [ 505.777474] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] result = hub.switch() [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return self.greenlet.switch() [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] result = function(*args, **kwargs) [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return func(*args, **kwargs) [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] raise e [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] nwinfo = self.network_api.allocate_for_instance( [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 505.777839] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] created_port_ids = self._update_ports_for_instance( [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] with excutils.save_and_reraise_exception(): [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self.force_reraise() [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] raise self.value [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] updated_port = self._update_port( [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] _ensure_no_port_binding_failure(port) [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 505.778247] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] raise exception.PortBindingFailed(port_id=port['id']) [ 505.778583] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] nova.exception.PortBindingFailed: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. [ 505.778583] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] [ 505.778583] env[61768]: INFO nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Terminating instance [ 505.783153] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 505.783396] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquired lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 505.783669] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 505.785960] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3b0c65-f73d-4d9a-9ce7-eb52fd09dc90 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.822138] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e20efb5-1257-47fc-b1a2-ccbd206f4f59 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.831229] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c3cc51-b970-4101-8e27-dc9c22e5daf4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 505.850243] env[61768]: DEBUG nova.compute.provider_tree [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 506.037416] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228511, 'name': CreateVM_Task, 'duration_secs': 1.294289} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 506.037654] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 506.038979] env[61768]: DEBUG oslo_vmware.service [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fdc90a-d611-491a-b22a-4692690a7364 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.046154] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 506.046359] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 506.047511] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 506.047511] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53054b5e-3d44-4118-9427-6a8ce38c6264 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.052651] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 506.052651] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525de988-ea93-b3e1-109b-0b2519b52278" [ 506.052651] env[61768]: _type = "Task" [ 506.052651] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.062307] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525de988-ea93-b3e1-109b-0b2519b52278, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.142542] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 506.142542] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 506.142542] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 506.142542] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 506.142542] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 506.142542] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 506.143153] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Didn't find any instances for network info cache update. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 506.143153] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 506.143153] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 506.147059] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 506.147297] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 506.147777] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 506.147777] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 506.147857] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 506.147971] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 506.178473] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Successfully created port: 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 506.328417] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 506.355976] env[61768]: DEBUG nova.scheduler.client.report [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 506.436600] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 506.567481] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 506.568202] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 506.568202] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 506.568202] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 506.568907] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 506.568907] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c644d398-5792-47ed-890b-a853283fa665 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.580817] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 506.581000] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 506.581804] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cd0a9c-2148-4f9f-a8e7-597bb425f3c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.589311] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26039c61-184e-44d5-8bd4-c74a91a3c0fc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.597915] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 506.597915] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272292a-1584-19f6-d611-77891dd74df1" [ 506.597915] env[61768]: _type = "Task" [ 506.597915] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 506.608187] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 506.610634] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272292a-1584-19f6-d611-77891dd74df1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 506.645511] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 506.645757] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 506.645936] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 506.646317] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 506.646533] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 506.646705] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 506.647905] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 506.648536] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 506.648536] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 506.648536] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 506.648669] env[61768]: DEBUG nova.virt.hardware [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 506.652028] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-489dc92e-cbbc-48e9-a70d-e8f85f78660e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.654126] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.660242] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39602461-07b5-4c6b-8354-1de8b0e9850c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.791728] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Acquiring lock "7d0e752f-9b39-4d01-bedc-5315927deea8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 506.792973] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Lock "7d0e752f-9b39-4d01-bedc-5315927deea8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.862029] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 506.862029] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 506.864054] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 3.151s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 506.939376] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Releasing lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 506.940187] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 506.940582] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 506.941019] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9476a98a-0663-4352-bf32-b7eb0f57d51d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.951584] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9ccbb0-8583-4537-8fde-4956887ec69e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 506.976147] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance df682d97-53c1-4a18-8800-b23f3fc66018 could not be found. [ 506.977432] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 506.977432] env[61768]: INFO nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Took 0.04 seconds to destroy the instance on the hypervisor. [ 506.977432] env[61768]: DEBUG oslo.service.loopingcall [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 506.977432] env[61768]: DEBUG nova.compute.manager [-] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 506.977432] env[61768]: DEBUG nova.network.neutron [-] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 506.997028] env[61768]: DEBUG nova.network.neutron [-] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 507.112732] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Preparing fetch location {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 507.113080] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Creating directory with path [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 507.113303] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e30f930e-6712-4d8d-b1a9-aeada3ae07c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.126127] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Created directory with path [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 507.126333] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Fetch image to [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 507.127071] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Downloading image file data 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk on the data store datastore2 {{(pid=61768) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 507.127925] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65258ee3-bf02-4485-b60d-55eb7bfad8c7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.135190] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e70cdc2-f182-4155-96ad-d88b758d5c17 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.151300] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01092597-954a-4c5a-84b2-3e43af6d6ce4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.186249] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d3d6c4-7194-4be5-ad0c-d54d1f97f2e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.192305] env[61768]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-8541d169-9639-41b2-88a1-e8897dcd62b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.222418] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Downloading image file data 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to the data store datastore2 {{(pid=61768) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 507.297092] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 507.315017] env[61768]: DEBUG oslo_vmware.rw_handles [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61768) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 507.387268] env[61768]: DEBUG nova.compute.utils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 507.396903] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 507.397153] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 507.503755] env[61768]: DEBUG nova.network.neutron [-] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 507.545253] env[61768]: DEBUG nova.policy [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '334f79403d0c4716846f8f8e1180ac4f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '124adecf56654cfcafc91bf4d3ae5ed2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 507.692027] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11a1f0e-7cd5-4806-868a-2531454c2ecd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.703319] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1e4cc3c-fe6a-4a24-aa94-4cd80690b15e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.759098] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f816a7-42da-4f52-83ec-5492f781b662 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.769050] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2069f861-b442-4efd-aa07-b7e0d821cf39 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 507.790105] env[61768]: DEBUG nova.compute.provider_tree [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 507.828188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 507.899744] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 508.007311] env[61768]: INFO nova.compute.manager [-] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Took 1.03 seconds to deallocate network for instance. [ 508.016804] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Acquiring lock "1b694ac3-385b-4083-a35c-542853354202" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.017104] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Lock "1b694ac3-385b-4083-a35c-542853354202" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.021573] env[61768]: DEBUG nova.compute.claims [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 508.021573] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 508.107900] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Successfully created port: 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 508.174843] env[61768]: DEBUG oslo_vmware.rw_handles [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Completed reading data from the image iterator. {{(pid=61768) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 508.174843] env[61768]: DEBUG oslo_vmware.rw_handles [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 508.232442] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Downloaded image file data 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk on the data store datastore2 {{(pid=61768) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 508.237346] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Caching image {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 508.237644] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Copying Virtual Disk [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk to [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 508.238518] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b3d99132-e793-4bfb-8505-6e7085c39f5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.246633] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 508.246633] env[61768]: value = "task-1228512" [ 508.246633] env[61768]: _type = "Task" [ 508.246633] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 508.258536] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.264117] env[61768]: ERROR nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. [ 508.264117] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 508.264117] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 508.264117] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 508.264117] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 508.264117] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 508.264117] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 508.264117] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 508.264117] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 508.264117] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 508.264117] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 508.264117] env[61768]: ERROR nova.compute.manager raise self.value [ 508.264117] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 508.264117] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 508.264117] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 508.264117] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 508.264907] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 508.264907] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 508.264907] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. [ 508.264907] env[61768]: ERROR nova.compute.manager [ 508.264907] env[61768]: Traceback (most recent call last): [ 508.264907] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 508.264907] env[61768]: listener.cb(fileno) [ 508.264907] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 508.264907] env[61768]: result = function(*args, **kwargs) [ 508.264907] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 508.264907] env[61768]: return func(*args, **kwargs) [ 508.264907] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 508.264907] env[61768]: raise e [ 508.264907] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 508.264907] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 508.264907] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 508.264907] env[61768]: created_port_ids = self._update_ports_for_instance( [ 508.264907] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 508.264907] env[61768]: with excutils.save_and_reraise_exception(): [ 508.264907] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 508.264907] env[61768]: self.force_reraise() [ 508.264907] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 508.264907] env[61768]: raise self.value [ 508.264907] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 508.264907] env[61768]: updated_port = self._update_port( [ 508.264907] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 508.264907] env[61768]: _ensure_no_port_binding_failure(port) [ 508.264907] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 508.264907] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 508.266674] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. [ 508.266674] env[61768]: Removing descriptor: 14 [ 508.266674] env[61768]: ERROR nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Traceback (most recent call last): [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] yield resources [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self.driver.spawn(context, instance, image_meta, [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self._vmops.spawn(context, instance, image_meta, injected_files, [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 508.266674] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] vm_ref = self.build_virtual_machine(instance, [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] vif_infos = vmwarevif.get_vif_info(self._session, [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] for vif in network_info: [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return self._sync_wrapper(fn, *args, **kwargs) [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self.wait() [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self[:] = self._gt.wait() [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return self._exit_event.wait() [ 508.267896] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] result = hub.switch() [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return self.greenlet.switch() [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] result = function(*args, **kwargs) [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return func(*args, **kwargs) [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] raise e [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] nwinfo = self.network_api.allocate_for_instance( [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 508.268431] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] created_port_ids = self._update_ports_for_instance( [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] with excutils.save_and_reraise_exception(): [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self.force_reraise() [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] raise self.value [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] updated_port = self._update_port( [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] _ensure_no_port_binding_failure(port) [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 508.268768] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] raise exception.PortBindingFailed(port_id=port['id']) [ 508.270689] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] nova.exception.PortBindingFailed: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. [ 508.270689] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] [ 508.270689] env[61768]: INFO nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Terminating instance [ 508.270689] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.270689] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquired lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.270689] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 508.274160] env[61768]: DEBUG nova.compute.manager [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Received event network-changed-c3e98df6-99ea-47a5-81e3-0fb22025e494 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 508.274160] env[61768]: DEBUG nova.compute.manager [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Refreshing instance network info cache due to event network-changed-c3e98df6-99ea-47a5-81e3-0fb22025e494. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 508.274160] env[61768]: DEBUG oslo_concurrency.lockutils [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] Acquiring lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.274160] env[61768]: DEBUG oslo_concurrency.lockutils [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] Acquired lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.274160] env[61768]: DEBUG nova.network.neutron [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Refreshing network info cache for port c3e98df6-99ea-47a5-81e3-0fb22025e494 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 508.299020] env[61768]: DEBUG nova.scheduler.client.report [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 508.520683] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 508.760252] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228512, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 508.806022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.939s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 508.806022] env[61768]: ERROR nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. [ 508.806022] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Traceback (most recent call last): [ 508.806022] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 508.806022] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self.driver.spawn(context, instance, image_meta, [ 508.806022] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 508.806022] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 508.806022] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 508.806022] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] vm_ref = self.build_virtual_machine(instance, [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] vif_infos = vmwarevif.get_vif_info(self._session, [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] for vif in network_info: [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return self._sync_wrapper(fn, *args, **kwargs) [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self.wait() [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self[:] = self._gt.wait() [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return self._exit_event.wait() [ 508.806426] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] result = hub.switch() [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return self.greenlet.switch() [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] result = function(*args, **kwargs) [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] return func(*args, **kwargs) [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] raise e [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] nwinfo = self.network_api.allocate_for_instance( [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 508.806761] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] created_port_ids = self._update_ports_for_instance( [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] with excutils.save_and_reraise_exception(): [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] self.force_reraise() [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] raise self.value [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] updated_port = self._update_port( [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] _ensure_no_port_binding_failure(port) [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 508.807100] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] raise exception.PortBindingFailed(port_id=port['id']) [ 508.807428] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] nova.exception.PortBindingFailed: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. [ 508.807428] env[61768]: ERROR nova.compute.manager [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] [ 508.807428] env[61768]: DEBUG nova.compute.utils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 508.812064] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.778s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 508.814437] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Build of instance 1efe0c6f-38e1-4131-9adc-995cbf33fa0c was re-scheduled: Binding failed for port 6f476265-1e10-4150-83d5-d936bb9146dc, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 508.815298] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 508.815775] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Acquiring lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 508.816786] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Acquired lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 508.816786] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 508.818809] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 508.848674] env[61768]: DEBUG nova.network.neutron [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 508.913270] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 508.943598] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 508.944175] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 508.944175] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 508.945048] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 508.945048] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 508.945048] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 508.945235] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 508.945304] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 508.945523] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 508.945627] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 508.946141] env[61768]: DEBUG nova.virt.hardware [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 508.946773] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3c751b-d44c-4ecb-a4af-9f0c674eb0c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.957191] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84240a3-4aa4-4719-833f-deefeb186045 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 508.996620] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.045097] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.166177] env[61768]: DEBUG nova.network.neutron [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.258783] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228512, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651284} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.259404] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Copied Virtual Disk [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk to [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 509.259741] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Deleting the datastore file [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 509.260110] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3b6fd7ec-9b2a-40c7-aa61-070ec43d2306 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.266581] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 509.266581] env[61768]: value = "task-1228513" [ 509.266581] env[61768]: _type = "Task" [ 509.266581] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.277065] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.387538] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 509.476428] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed216b49-f1f4-4810-ab2c-f61016fa3f95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.490041] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0325f31-7dc7-4e74-aba6-021d1b5c5503 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.522672] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Releasing lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 509.523174] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 509.523392] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 509.524195] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-47519224-e3e9-4181-b1a9-1ea0231ea0fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.526624] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-464d0efb-c822-4c4c-b386-d38f5e7a210c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.540291] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249fe997-037a-46a4-93d6-ceacc65dc74b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.556352] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6be9df-9f8e-4936-ad02-e286917bccf0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.567187] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 438d040b-c86f-47ab-8d22-e0e86c41ac43 could not be found. [ 509.571902] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 509.571902] env[61768]: INFO nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Took 0.05 seconds to destroy the instance on the hypervisor. [ 509.571902] env[61768]: DEBUG oslo.service.loopingcall [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 509.580149] env[61768]: DEBUG nova.compute.manager [-] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 509.580149] env[61768]: DEBUG nova.network.neutron [-] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 509.581988] env[61768]: DEBUG nova.compute.provider_tree [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 509.584389] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 509.620998] env[61768]: DEBUG nova.network.neutron [-] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 509.668472] env[61768]: DEBUG oslo_concurrency.lockutils [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] Releasing lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 509.668734] env[61768]: DEBUG nova.compute.manager [req-6123da03-c941-4aed-ba25-12145924399b req-50243256-3c27-4647-9135-cfdafa0a8fff service nova] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Received event network-vif-deleted-c3e98df6-99ea-47a5-81e3-0fb22025e494 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 509.777901] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042254} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 509.778700] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 509.778700] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Moving file from [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c/0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354. {{(pid=61768) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 509.778852] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-355ee30e-0b2f-487e-8e70-615d988b0fb1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 509.785729] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 509.785729] env[61768]: value = "task-1228514" [ 509.785729] env[61768]: _type = "Task" [ 509.785729] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 509.794218] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228514, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 509.919234] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "417480ad-61d9-447d-a84d-0cf5df94f4bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 509.919413] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "417480ad-61d9-447d-a84d-0cf5df94f4bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.085578] env[61768]: DEBUG nova.scheduler.client.report [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 510.091739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Releasing lock "refresh_cache-1efe0c6f-38e1-4131-9adc-995cbf33fa0c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.091856] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 510.091968] env[61768]: DEBUG nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 510.092778] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 510.124677] env[61768]: DEBUG nova.network.neutron [-] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 510.132146] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 510.300998] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228514, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.026748} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.301399] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] File moved {{(pid=61768) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 510.301599] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Cleaning up location [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 510.301759] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Deleting the datastore file [datastore2] vmware_temp/bddd91c2-d5ed-4078-9993-b0233411912c {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 510.302048] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-18589748-c3eb-43f5-a455-4c2ebe754393 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.309389] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 510.309389] env[61768]: value = "task-1228515" [ 510.309389] env[61768]: _type = "Task" [ 510.309389] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.325044] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228515, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 510.475709] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "eb8203da-3427-48a8-9ec0-a9d12d8c44d2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.475709] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "eb8203da-3427-48a8-9ec0-a9d12d8c44d2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.595026] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.783s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 510.595026] env[61768]: ERROR nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. [ 510.595026] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Traceback (most recent call last): [ 510.595026] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 510.595026] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self.driver.spawn(context, instance, image_meta, [ 510.595026] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 510.595026] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 510.595026] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 510.595026] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] vm_ref = self.build_virtual_machine(instance, [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] vif_infos = vmwarevif.get_vif_info(self._session, [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] for vif in network_info: [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return self._sync_wrapper(fn, *args, **kwargs) [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self.wait() [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self[:] = self._gt.wait() [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return self._exit_event.wait() [ 510.595460] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] result = hub.switch() [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return self.greenlet.switch() [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] result = function(*args, **kwargs) [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] return func(*args, **kwargs) [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] raise e [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] nwinfo = self.network_api.allocate_for_instance( [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 510.595937] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] created_port_ids = self._update_ports_for_instance( [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] with excutils.save_and_reraise_exception(): [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] self.force_reraise() [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] raise self.value [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] updated_port = self._update_port( [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] _ensure_no_port_binding_failure(port) [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 510.596413] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] raise exception.PortBindingFailed(port_id=port['id']) [ 510.596807] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] nova.exception.PortBindingFailed: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. [ 510.596807] env[61768]: ERROR nova.compute.manager [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] [ 510.596807] env[61768]: DEBUG nova.compute.utils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 510.599828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.285s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.604369] env[61768]: INFO nova.compute.claims [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 510.609621] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Build of instance 0b1de80e-6342-4443-a700-3d50f28270b6 was re-scheduled: Binding failed for port 8fe593a8-1b08-4764-8ba1-e45d2658db24, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 510.609621] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 510.609621] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Acquiring lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 510.609817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Acquired lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 510.610457] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 510.630250] env[61768]: INFO nova.compute.manager [-] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Took 1.05 seconds to deallocate network for instance. [ 510.630972] env[61768]: DEBUG nova.compute.claims [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 510.632988] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.634662] env[61768]: DEBUG nova.network.neutron [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 510.828539] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228515, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042368} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.828795] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 510.829577] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db378887-7c27-48f6-ab5f-bd108dfe7aec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.838776] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 510.838776] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5284a173-d8cf-758f-50b1-d694f01fbaba" [ 510.838776] env[61768]: _type = "Task" [ 510.838776] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.848688] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5284a173-d8cf-758f-50b1-d694f01fbaba, 'name': SearchDatastore_Task, 'duration_secs': 0.008632} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 510.848688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 510.848947] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] ca465e62-c190-4d67-9dec-87ea1aa8f194/ca465e62-c190-4d67-9dec-87ea1aa8f194.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 510.849222] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8bee4009-a53c-4baa-8a82-408ce98357a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.856999] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 510.856999] env[61768]: value = "task-1228516" [ 510.856999] env[61768]: _type = "Task" [ 510.856999] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 510.869790] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "7bc2075c-2cb2-41c2-9681-780689c1d080" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.869987] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "7bc2075c-2cb2-41c2-9681-780689c1d080" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 510.876161] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228516, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.045810] env[61768]: DEBUG nova.compute.manager [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Received event network-changed-5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 511.045916] env[61768]: DEBUG nova.compute.manager [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Refreshing instance network info cache due to event network-changed-5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 511.047692] env[61768]: DEBUG oslo_concurrency.lockutils [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] Acquiring lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 511.048231] env[61768]: DEBUG oslo_concurrency.lockutils [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] Acquired lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 511.050398] env[61768]: DEBUG nova.network.neutron [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Refreshing network info cache for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 511.137301] env[61768]: INFO nova.compute.manager [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] [instance: 1efe0c6f-38e1-4131-9adc-995cbf33fa0c] Took 1.04 seconds to deallocate network for instance. [ 511.241215] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 511.372292] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228516, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450509} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.373371] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] ca465e62-c190-4d67-9dec-87ea1aa8f194/ca465e62-c190-4d67-9dec-87ea1aa8f194.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 511.373592] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 511.373844] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-79e1564a-88af-4470-9851-e6f18e47f090 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.382558] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 511.382558] env[61768]: value = "task-1228517" [ 511.382558] env[61768]: _type = "Task" [ 511.382558] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 511.396757] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228517, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 511.428236] env[61768]: ERROR nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. [ 511.428236] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 511.428236] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 511.428236] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 511.428236] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 511.428236] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 511.428236] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 511.428236] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 511.428236] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.428236] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 511.428236] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.428236] env[61768]: ERROR nova.compute.manager raise self.value [ 511.428236] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 511.428236] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 511.428236] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.428236] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 511.428853] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.428853] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 511.428853] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. [ 511.428853] env[61768]: ERROR nova.compute.manager [ 511.428853] env[61768]: Traceback (most recent call last): [ 511.428853] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 511.428853] env[61768]: listener.cb(fileno) [ 511.428853] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 511.428853] env[61768]: result = function(*args, **kwargs) [ 511.428853] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 511.428853] env[61768]: return func(*args, **kwargs) [ 511.428853] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 511.428853] env[61768]: raise e [ 511.428853] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 511.428853] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 511.428853] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 511.428853] env[61768]: created_port_ids = self._update_ports_for_instance( [ 511.428853] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 511.428853] env[61768]: with excutils.save_and_reraise_exception(): [ 511.428853] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.428853] env[61768]: self.force_reraise() [ 511.428853] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.428853] env[61768]: raise self.value [ 511.428853] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 511.428853] env[61768]: updated_port = self._update_port( [ 511.428853] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.428853] env[61768]: _ensure_no_port_binding_failure(port) [ 511.428853] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.428853] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 511.430154] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. [ 511.430154] env[61768]: Removing descriptor: 17 [ 511.430154] env[61768]: ERROR nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Traceback (most recent call last): [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] yield resources [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self.driver.spawn(context, instance, image_meta, [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self._vmops.spawn(context, instance, image_meta, injected_files, [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 511.430154] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] vm_ref = self.build_virtual_machine(instance, [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] vif_infos = vmwarevif.get_vif_info(self._session, [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] for vif in network_info: [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return self._sync_wrapper(fn, *args, **kwargs) [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self.wait() [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self[:] = self._gt.wait() [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return self._exit_event.wait() [ 511.430990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] result = hub.switch() [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return self.greenlet.switch() [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] result = function(*args, **kwargs) [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return func(*args, **kwargs) [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] raise e [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] nwinfo = self.network_api.allocate_for_instance( [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 511.431383] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] created_port_ids = self._update_ports_for_instance( [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] with excutils.save_and_reraise_exception(): [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self.force_reraise() [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] raise self.value [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] updated_port = self._update_port( [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] _ensure_no_port_binding_failure(port) [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 511.431785] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] raise exception.PortBindingFailed(port_id=port['id']) [ 511.432151] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] nova.exception.PortBindingFailed: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. [ 511.432151] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] [ 511.432151] env[61768]: INFO nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Terminating instance [ 511.432240] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Acquiring lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 511.432559] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Acquired lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 511.432667] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 511.519732] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.575103] env[61768]: DEBUG nova.network.neutron [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 511.737462] env[61768]: DEBUG nova.network.neutron [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 511.886875] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-978fd832-2b42-4d04-8fe1-5b8190d452a5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.902305] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e3649c-ffd7-4da7-af90-d23b275afb15 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.906706] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228517, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06447} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 511.907085] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 511.908661] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2185351e-026e-4fcd-8ea8-25bb66477072 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.939450] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f4f59d-46f3-452d-b462-91dd11c75e64 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.960150] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] ca465e62-c190-4d67-9dec-87ea1aa8f194/ca465e62-c190-4d67-9dec-87ea1aa8f194.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 511.960875] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-277b4336-8731-4255-a46a-26bce2025152 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.981387] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 511.985379] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a994662-4a6c-4d25-8c48-bbd5cc73faf1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 511.991804] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 511.991804] env[61768]: value = "task-1228518" [ 511.991804] env[61768]: _type = "Task" [ 511.991804] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.005680] env[61768]: DEBUG nova.compute.provider_tree [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 512.013518] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228518, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.023113] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Releasing lock "refresh_cache-0b1de80e-6342-4443-a700-3d50f28270b6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.023113] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 512.023113] env[61768]: DEBUG nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 512.023113] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 512.049419] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 512.185751] env[61768]: INFO nova.scheduler.client.report [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Deleted allocations for instance 1efe0c6f-38e1-4131-9adc-995cbf33fa0c [ 512.244784] env[61768]: DEBUG oslo_concurrency.lockutils [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] Releasing lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.244784] env[61768]: DEBUG nova.compute.manager [req-9da2a4fe-56cb-4bde-9bf6-09eaf98b4156 req-0367d554-df42-44ac-98f3-e1b7cd837806 service nova] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Received event network-vif-deleted-5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 512.415030] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 512.506698] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228518, 'name': ReconfigVM_Task, 'duration_secs': 0.285578} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 512.507418] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Reconfigured VM instance instance-00000004 to attach disk [datastore2] ca465e62-c190-4d67-9dec-87ea1aa8f194/ca465e62-c190-4d67-9dec-87ea1aa8f194.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 512.508064] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87037196-d4eb-402c-9971-8d1ad52065de {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.510261] env[61768]: DEBUG nova.scheduler.client.report [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 512.520937] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 512.520937] env[61768]: value = "task-1228519" [ 512.520937] env[61768]: _type = "Task" [ 512.520937] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 512.531289] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228519, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 512.554077] env[61768]: DEBUG nova.network.neutron [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 512.702701] env[61768]: DEBUG oslo_concurrency.lockutils [None req-97b1cef5-8b45-4aba-9008-776bd2c7d4cf tempest-ServerDiagnosticsTest-526574239 tempest-ServerDiagnosticsTest-526574239-project-member] Lock "1efe0c6f-38e1-4131-9adc-995cbf33fa0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.685s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 512.920232] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Releasing lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 512.920712] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 512.920835] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 512.921261] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-141fa34e-1aa2-497a-a55c-a8ee699ae119 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.932374] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d2159e-ba6f-4611-aad9-6315645be3d7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.961292] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0188dbb6-ae91-4601-83a3-3e3bf55fc753 could not be found. [ 512.961292] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 512.961292] env[61768]: INFO nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Took 0.04 seconds to destroy the instance on the hypervisor. [ 512.961292] env[61768]: DEBUG oslo.service.loopingcall [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 512.961292] env[61768]: DEBUG nova.compute.manager [-] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 512.961292] env[61768]: DEBUG nova.network.neutron [-] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 512.991182] env[61768]: DEBUG nova.network.neutron [-] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 513.000717] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Acquiring lock "ca31b2ec-0bfb-47d9-9f35-c54659fa7799" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.001610] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Lock "ca31b2ec-0bfb-47d9-9f35-c54659fa7799" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.015792] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.421s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.017472] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 513.023747] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.295s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.023747] env[61768]: INFO nova.compute.claims [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 513.042390] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228519, 'name': Rename_Task, 'duration_secs': 0.127594} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.042390] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 513.042390] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1768cfb4-fb63-42a6-adcf-cc096f847f33 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.046831] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 513.046831] env[61768]: value = "task-1228520" [ 513.046831] env[61768]: _type = "Task" [ 513.046831] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 513.060551] env[61768]: INFO nova.compute.manager [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] [instance: 0b1de80e-6342-4443-a700-3d50f28270b6] Took 1.04 seconds to deallocate network for instance. [ 513.064379] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228520, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 513.208778] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 513.493325] env[61768]: DEBUG nova.network.neutron [-] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 513.530621] env[61768]: DEBUG nova.compute.utils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 513.541190] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 513.541378] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 513.561458] env[61768]: DEBUG oslo_vmware.api [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228520, 'name': PowerOnVM_Task, 'duration_secs': 0.430391} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 513.561690] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 513.561772] env[61768]: INFO nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Took 9.20 seconds to spawn the instance on the hypervisor. [ 513.562028] env[61768]: DEBUG nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 513.562849] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec74c74e-e33b-4db4-905d-f8d149b6fce0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 513.617575] env[61768]: DEBUG nova.policy [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a095b6d033e649c0a60f572c3893f95d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee67e53b0b544fda9ff343cff7a802f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 513.750461] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.995910] env[61768]: INFO nova.compute.manager [-] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Took 1.04 seconds to deallocate network for instance. [ 514.001821] env[61768]: DEBUG nova.compute.claims [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 514.002915] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.023820] env[61768]: DEBUG nova.compute.manager [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Received event network-changed-44ca2d9d-2ce5-47ed-9e5b-8b349b46e524 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 514.023820] env[61768]: DEBUG nova.compute.manager [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Refreshing instance network info cache due to event network-changed-44ca2d9d-2ce5-47ed-9e5b-8b349b46e524. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 514.023820] env[61768]: DEBUG oslo_concurrency.lockutils [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] Acquiring lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 514.023820] env[61768]: DEBUG oslo_concurrency.lockutils [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] Acquired lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 514.023820] env[61768]: DEBUG nova.network.neutron [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Refreshing network info cache for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 514.043032] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 514.090880] env[61768]: INFO nova.compute.manager [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Took 17.91 seconds to build instance. [ 514.115747] env[61768]: INFO nova.scheduler.client.report [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Deleted allocations for instance 0b1de80e-6342-4443-a700-3d50f28270b6 [ 514.328442] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1bbef8-4b01-4727-8b60-c643cf8c9a0c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.336204] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce67692-d4d1-4177-b132-7c93203ffaf9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.368067] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Successfully created port: 9fbb9c99-fa97-43b0-a4b0-6774ff75b899 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 514.370488] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fcefbf-e729-4a74-88f7-7dd3d1d96134 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.378105] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99c35c5-cc01-445e-9ed0-8f0407c98217 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 514.392395] env[61768]: DEBUG nova.compute.provider_tree [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 514.558596] env[61768]: DEBUG nova.network.neutron [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 514.593426] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f9a2a882-d933-438f-9c2d-04d0e6a1182f tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "ca465e62-c190-4d67-9dec-87ea1aa8f194" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.419s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 514.628882] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1c1da65a-9d38-45e8-9958-52e2cb1ac3f8 tempest-ServersAdminNegativeTestJSON-331283824 tempest-ServersAdminNegativeTestJSON-331283824-project-member] Lock "0b1de80e-6342-4443-a700-3d50f28270b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.007s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 514.754939] env[61768]: DEBUG nova.network.neutron [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 514.896620] env[61768]: DEBUG nova.scheduler.client.report [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 515.062700] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 515.096627] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.096627] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.096627] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.097177] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.097177] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.097177] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.097177] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.097177] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.097334] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.097368] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.097568] env[61768]: DEBUG nova.virt.hardware [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.097964] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.101734] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1df7c47-7034-4e93-a3f4-c585a3ec2d56 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.113712] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d56fa804-388c-4d8b-b9de-671ea51ea17b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.135722] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.258284] env[61768]: DEBUG oslo_concurrency.lockutils [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] Releasing lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 515.259069] env[61768]: DEBUG nova.compute.manager [req-6a6129c0-a18c-4531-8036-08700e669bb5 req-b834a354-9a87-4a62-ac8f-84c6be14e5fa service nova] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Received event network-vif-deleted-44ca2d9d-2ce5-47ed-9e5b-8b349b46e524 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 515.406656] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.406656] env[61768]: DEBUG nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 515.411084] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.757s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.411399] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 515.412294] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 515.412640] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.585s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.415366] env[61768]: INFO nova.compute.claims [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.422472] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fcab5a9-f048-4e73-8c16-948e39935325 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.434388] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e0c6bc-064d-4513-b4aa-ff5d0c2956a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.462302] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beac8820-d445-4ac3-9525-6465d945a840 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.471802] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28463056-6e5f-47fc-a240-0c90cd20dea1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.506568] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181505MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 515.506568] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.635140] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.668931] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.927230] env[61768]: DEBUG nova.compute.utils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 515.928911] env[61768]: DEBUG nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Not allocating networking since 'none' was specified. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 516.435207] env[61768]: DEBUG nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 516.588201] env[61768]: DEBUG nova.compute.manager [None req-ef460549-1d87-406b-b5dc-d89bc05add20 tempest-ServerDiagnosticsV248Test-1840855018 tempest-ServerDiagnosticsV248Test-1840855018-project-admin] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 516.588546] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66999798-874d-4248-929f-567029929fe0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.600117] env[61768]: INFO nova.compute.manager [None req-ef460549-1d87-406b-b5dc-d89bc05add20 tempest-ServerDiagnosticsV248Test-1840855018 tempest-ServerDiagnosticsV248Test-1840855018-project-admin] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Retrieving diagnostics [ 516.600996] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e5cdf16-de9f-44f3-859e-938456a4a62e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.674984] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51965bdc-04f2-4fee-afb6-50314d714123 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.684639] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e3ff3f-4237-4919-a5c3-fef8ffd95e9a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.720971] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb5f4b8-0197-48c5-9c86-b3a4cd92063c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.729373] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81bcc40-d27d-4e5a-ad69-25a6639b22b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.748976] env[61768]: DEBUG nova.compute.provider_tree [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.113342] env[61768]: ERROR nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. [ 517.113342] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 517.113342] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 517.113342] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 517.113342] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 517.113342] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 517.113342] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 517.113342] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 517.113342] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 517.113342] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 517.113342] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 517.113342] env[61768]: ERROR nova.compute.manager raise self.value [ 517.113342] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 517.113342] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 517.113342] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 517.113342] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 517.113891] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 517.113891] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 517.113891] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. [ 517.113891] env[61768]: ERROR nova.compute.manager [ 517.113891] env[61768]: Traceback (most recent call last): [ 517.113891] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 517.113891] env[61768]: listener.cb(fileno) [ 517.113891] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 517.113891] env[61768]: result = function(*args, **kwargs) [ 517.113891] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 517.113891] env[61768]: return func(*args, **kwargs) [ 517.113891] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 517.113891] env[61768]: raise e [ 517.113891] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 517.113891] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 517.113891] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 517.113891] env[61768]: created_port_ids = self._update_ports_for_instance( [ 517.113891] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 517.113891] env[61768]: with excutils.save_and_reraise_exception(): [ 517.113891] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 517.113891] env[61768]: self.force_reraise() [ 517.113891] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 517.113891] env[61768]: raise self.value [ 517.113891] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 517.113891] env[61768]: updated_port = self._update_port( [ 517.113891] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 517.113891] env[61768]: _ensure_no_port_binding_failure(port) [ 517.113891] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 517.113891] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 517.114700] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. [ 517.114700] env[61768]: Removing descriptor: 14 [ 517.114700] env[61768]: ERROR nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Traceback (most recent call last): [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] yield resources [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self.driver.spawn(context, instance, image_meta, [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 517.114700] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] vm_ref = self.build_virtual_machine(instance, [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] for vif in network_info: [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return self._sync_wrapper(fn, *args, **kwargs) [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self.wait() [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self[:] = self._gt.wait() [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return self._exit_event.wait() [ 517.115184] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] result = hub.switch() [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return self.greenlet.switch() [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] result = function(*args, **kwargs) [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return func(*args, **kwargs) [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] raise e [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] nwinfo = self.network_api.allocate_for_instance( [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 517.115842] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] created_port_ids = self._update_ports_for_instance( [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] with excutils.save_and_reraise_exception(): [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self.force_reraise() [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] raise self.value [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] updated_port = self._update_port( [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] _ensure_no_port_binding_failure(port) [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 517.116267] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] raise exception.PortBindingFailed(port_id=port['id']) [ 517.116602] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] nova.exception.PortBindingFailed: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. [ 517.116602] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] [ 517.116602] env[61768]: INFO nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Terminating instance [ 517.118307] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 517.119094] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquired lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 517.119094] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 517.251206] env[61768]: DEBUG nova.scheduler.client.report [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.450272] env[61768]: DEBUG nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 517.479318] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 517.479551] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 517.479709] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 517.479930] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 517.480489] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 517.480783] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 517.481126] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 517.481457] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 517.481747] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 517.481905] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 517.482206] env[61768]: DEBUG nova.virt.hardware [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 517.483296] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16db62a3-c35a-4c90-aa64-d96260e33da3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.494715] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39fc9281-9084-47de-ab07-b70997a8db4d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.509789] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 517.516990] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Creating folder: Project (45380e3cf1b74cc19b263eb6ef8cf372). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 517.517338] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-14eb9d2e-6021-434b-bf32-3a191e795b95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.529666] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Created folder: Project (45380e3cf1b74cc19b263eb6ef8cf372) in parent group-v265360. [ 517.529779] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Creating folder: Instances. Parent ref: group-v265364. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 517.529999] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd8ae28c-f4a7-4207-973c-33b9564014d0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.538597] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Created folder: Instances in parent group-v265364. [ 517.538849] env[61768]: DEBUG oslo.service.loopingcall [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 517.539061] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 517.539385] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22236de0-9ea6-46d2-99d6-f5ad35416fc5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.559232] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 517.559232] env[61768]: value = "task-1228523" [ 517.559232] env[61768]: _type = "Task" [ 517.559232] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 517.567348] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228523, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 517.654941] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 517.759609] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 517.760163] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.763554] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.745s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.873062] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 517.907905] env[61768]: DEBUG nova.compute.manager [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Received event network-changed-9fbb9c99-fa97-43b0-a4b0-6774ff75b899 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 517.909264] env[61768]: DEBUG nova.compute.manager [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Refreshing instance network info cache due to event network-changed-9fbb9c99-fa97-43b0-a4b0-6774ff75b899. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 517.909264] env[61768]: DEBUG oslo_concurrency.lockutils [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] Acquiring lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.071801] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228523, 'name': CreateVM_Task, 'duration_secs': 0.408697} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 518.071971] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 518.072847] env[61768]: DEBUG oslo_vmware.service [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a099d177-7fe1-40bc-9604-a55b79be9afb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.081131] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.081131] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.081131] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 518.081131] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7de8df45-a608-43b2-ae12-6f7d351b4d04 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.088454] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 518.088454] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5244f085-2aa9-fc5e-8c99-91cac00e342d" [ 518.088454] env[61768]: _type = "Task" [ 518.088454] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.098454] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5244f085-2aa9-fc5e-8c99-91cac00e342d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.271469] env[61768]: DEBUG nova.compute.utils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.275896] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 518.276083] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 518.377930] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Releasing lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 518.378693] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 518.378921] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 518.383543] env[61768]: DEBUG oslo_concurrency.lockutils [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] Acquired lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.383543] env[61768]: DEBUG nova.network.neutron [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Refreshing network info cache for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 518.384713] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8d42f6e3-6480-4244-b470-6233633a829b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.394420] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82afc6a3-a0ad-488d-b163-e0f10f9cfdc7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.420514] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c3b90a4d-e885-453c-9f8e-799adb2400a0 could not be found. [ 518.420732] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 518.420913] env[61768]: INFO nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 518.421169] env[61768]: DEBUG oslo.service.loopingcall [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 518.424037] env[61768]: DEBUG nova.compute.manager [-] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 518.424037] env[61768]: DEBUG nova.network.neutron [-] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 518.480667] env[61768]: DEBUG nova.network.neutron [-] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 518.541377] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0037b2e1-47dd-489e-be4d-d91106f21d09 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.550677] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c14e88-24c7-4ce2-8924-80e4c6dcab21 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.586234] env[61768]: DEBUG nova.policy [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c6e3b642928497a9c5c3a532c70049d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e8526bd9ab9a4a61adc36ab8a9ddb014', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 518.590561] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e15169-7aa8-495d-8b4e-2c20cd46cf56 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.607731] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2c297f-8a4b-423f-b9db-b1689f632d9c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.611101] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 518.612194] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 518.612194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 518.612194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 518.612194] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 518.612808] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1520f7a-222b-4b1b-9051-fd690be1243e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.623625] env[61768]: DEBUG nova.compute.provider_tree [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 518.643371] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 518.643371] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 518.643371] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81ca059-aa36-45c9-8203-e7a792156adc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.652419] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-975219ea-1d0f-42e8-8dee-9d798464bba2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.660165] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 518.660165] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527f9093-aa0c-6f28-d99d-987d22549fab" [ 518.660165] env[61768]: _type = "Task" [ 518.660165] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 518.668522] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527f9093-aa0c-6f28-d99d-987d22549fab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 518.674134] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Acquiring lock "7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.674134] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Lock "7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.778456] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 518.984167] env[61768]: DEBUG nova.network.neutron [-] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 518.988096] env[61768]: DEBUG nova.network.neutron [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 519.130523] env[61768]: DEBUG nova.scheduler.client.report [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 519.171710] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Preparing fetch location {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 519.172065] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Creating directory with path [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 519.172561] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5753998e-2de9-4eea-b5bf-e6308318e8b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.209503] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Created directory with path [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 519.209795] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Fetch image to [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 519.209885] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Downloading image file data 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk on the data store datastore1 {{(pid=61768) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 519.210691] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-236e8ffc-e017-4ab3-a7cf-bfae5b45f015 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.222246] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc06a3d8-8978-4205-b742-63f08f17dc31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.235617] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a065614-82f4-422a-8d0a-6202b415889c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.278895] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13cb387-1243-4406-bc9b-9cea4849c475 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.291332] env[61768]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-44bee000-1cf7-4173-9dc7-4d30261afd35 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.379294] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Downloading image file data 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to the data store datastore1 {{(pid=61768) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 519.441144] env[61768]: DEBUG oslo_vmware.rw_handles [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61768) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 519.505580] env[61768]: INFO nova.compute.manager [-] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Took 1.08 seconds to deallocate network for instance. [ 519.506745] env[61768]: DEBUG nova.network.neutron [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 519.514237] env[61768]: DEBUG nova.compute.claims [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 519.514463] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.636221] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.872s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 519.636900] env[61768]: ERROR nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Traceback (most recent call last): [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self.driver.spawn(context, instance, image_meta, [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self._vmops.spawn(context, instance, image_meta, injected_files, [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] vm_ref = self.build_virtual_machine(instance, [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] vif_infos = vmwarevif.get_vif_info(self._session, [ 519.636900] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] for vif in network_info: [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return self._sync_wrapper(fn, *args, **kwargs) [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self.wait() [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self[:] = self._gt.wait() [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return self._exit_event.wait() [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] result = hub.switch() [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 519.637343] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return self.greenlet.switch() [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] result = function(*args, **kwargs) [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] return func(*args, **kwargs) [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] raise e [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] nwinfo = self.network_api.allocate_for_instance( [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] created_port_ids = self._update_ports_for_instance( [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] with excutils.save_and_reraise_exception(): [ 519.637700] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] self.force_reraise() [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] raise self.value [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] updated_port = self._update_port( [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] _ensure_no_port_binding_failure(port) [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] raise exception.PortBindingFailed(port_id=port['id']) [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] nova.exception.PortBindingFailed: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. [ 519.638096] env[61768]: ERROR nova.compute.manager [instance: df682d97-53c1-4a18-8800-b23f3fc66018] [ 519.638609] env[61768]: DEBUG nova.compute.utils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 519.648316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.599s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.648316] env[61768]: INFO nova.compute.claims [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 519.657853] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Build of instance df682d97-53c1-4a18-8800-b23f3fc66018 was re-scheduled: Binding failed for port c3e98df6-99ea-47a5-81e3-0fb22025e494, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 519.658378] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 519.658609] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.658754] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquired lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.658912] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 519.789481] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 519.826688] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 519.826688] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 519.826688] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 519.826901] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 519.829127] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 519.829127] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 519.829127] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 519.829127] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 519.829127] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 519.830711] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 519.830711] env[61768]: DEBUG nova.virt.hardware [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 519.830711] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9232cb65-b204-4b85-8f69-7982fd9b1e3e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.846840] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025c38f6-3666-4b3b-b4dc-942ccb11e6f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.012186] env[61768]: DEBUG oslo_concurrency.lockutils [req-caa56c05-fd3d-4c61-a293-135a5b54a9b3 req-a479ce9a-f86b-481c-8468-7a115f64e5fd service nova] Releasing lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.257320] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 520.278473] env[61768]: DEBUG oslo_vmware.rw_handles [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Completed reading data from the image iterator. {{(pid=61768) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 520.278702] env[61768]: DEBUG oslo_vmware.rw_handles [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 520.346460] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Downloaded image file data 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk on the data store datastore1 {{(pid=61768) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 520.347555] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Caching image {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 520.348198] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Copying Virtual Disk [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk to [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 520.348198] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-db68d502-f427-4c7f-bbd3-cb507ba04f13 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.358245] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 520.358245] env[61768]: value = "task-1228524" [ 520.358245] env[61768]: _type = "Task" [ 520.358245] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 520.372310] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228524, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.475897] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.591673] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Successfully created port: 65e56e3c-761f-4ec3-994c-8b1f396119ed {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 520.711737] env[61768]: DEBUG nova.compute.manager [req-35c1e0d6-087e-419d-a6ae-024ebbdbef70 req-35a8d8b6-baa1-4eea-9568-843bb400d7ef service nova] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Received event network-vif-deleted-9fbb9c99-fa97-43b0-a4b0-6774ff75b899 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 520.833500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Acquiring lock "1331a447-7be8-42b8-9d59-6b97742e49d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.833993] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Lock "1331a447-7be8-42b8-9d59-6b97742e49d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.005s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.877785] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228524, 'name': CopyVirtualDisk_Task} progress is 27%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 520.985463] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Releasing lock "refresh_cache-df682d97-53c1-4a18-8800-b23f3fc66018" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.985720] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 520.985883] env[61768]: DEBUG nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 520.986078] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 521.013202] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3d0275-9f9f-4d44-bfcb-2b96b7e003aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.022142] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe71adc-5f40-4bc1-8866-805c7ddbf000 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.065306] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6077f425-b2b3-425f-ba5a-4ce5cd1a3555 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.073835] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fa2fa2d-e48c-494e-9f12-ba1bdf54137a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.078834] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 521.091340] env[61768]: DEBUG nova.compute.provider_tree [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.375034] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228524, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 521.582608] env[61768]: DEBUG nova.network.neutron [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.594155] env[61768]: DEBUG nova.scheduler.client.report [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 521.875328] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228524, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.031315} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 521.875995] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Copied Virtual Disk [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk to [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 521.876350] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleting the datastore file [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/tmp-sparse.vmdk {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 521.877085] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f2c9bb5c-ac41-4954-9bcb-ba6841653494 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.884658] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 521.884658] env[61768]: value = "task-1228525" [ 521.884658] env[61768]: _type = "Task" [ 521.884658] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 521.894047] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228525, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.087334] env[61768]: INFO nova.compute.manager [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: df682d97-53c1-4a18-8800-b23f3fc66018] Took 1.10 seconds to deallocate network for instance. [ 522.099519] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.100043] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.104146] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.473s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.401140] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228525, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023332} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.401140] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 522.403040] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Moving file from [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d/0a07405a-3ac2-4d7f-bd39-a1b659ab2354 to [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354. {{(pid=61768) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 522.403330] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-4aba7833-ff64-4294-9316-4ecb26fcd10e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.410879] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 522.410879] env[61768]: value = "task-1228526" [ 522.410879] env[61768]: _type = "Task" [ 522.410879] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.422837] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228526, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 522.606580] env[61768]: DEBUG nova.compute.utils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 522.608021] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 522.608255] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 522.816970] env[61768]: DEBUG nova.policy [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f587daad67ee477c8c38d5274792083a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2a88af825b494ea2993466ee197a4efd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 522.921854] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228526, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024367} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.923152] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] File moved {{(pid=61768) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 522.923302] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Cleaning up location [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 522.923512] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleting the datastore file [datastore1] vmware_temp/7fb1c610-3976-465e-8dc9-eaff8499e82d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 522.924270] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d118810-1df8-4d9d-a78c-81cb304a3f42 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.927370] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b5cd427-54b9-4f3d-8b0c-baebbac05598 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.935754] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529b7cba-1d9b-49bb-9c23-545b6677a58b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.939042] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 522.939042] env[61768]: value = "task-1228527" [ 522.939042] env[61768]: _type = "Task" [ 522.939042] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 522.969673] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10757e3c-6e3c-4e88-a12f-f55ed0791f32 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.975729] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228527, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025331} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 522.976421] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 522.977409] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a2a4e5f-f083-4b1f-9c78-823eef24fadd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.983364] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b0d363-9c5e-4a62-b470-541885a1968f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.988489] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 522.988489] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f861fd-1fb6-7672-0cbe-9a4db838d5b5" [ 522.988489] env[61768]: _type = "Task" [ 522.988489] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.000078] env[61768]: DEBUG nova.compute.provider_tree [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 523.006246] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f861fd-1fb6-7672-0cbe-9a4db838d5b5, 'name': SearchDatastore_Task, 'duration_secs': 0.007992} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.006515] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.006787] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 523.007878] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-23e33d7b-dc77-40a2-9efa-6309f756df26 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.012744] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 523.012744] env[61768]: value = "task-1228528" [ 523.012744] env[61768]: _type = "Task" [ 523.012744] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.021255] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228528, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.116435] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.131294] env[61768]: INFO nova.scheduler.client.report [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Deleted allocations for instance df682d97-53c1-4a18-8800-b23f3fc66018 [ 523.509037] env[61768]: DEBUG nova.scheduler.client.report [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 523.531912] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228528, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.45898} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 523.532198] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 523.532422] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 523.532674] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6fe5c562-27a4-4089-afab-835397bb31fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.542802] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 523.542802] env[61768]: value = "task-1228529" [ 523.542802] env[61768]: _type = "Task" [ 523.542802] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 523.559771] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228529, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 523.638944] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22ce9d17-0d91-4427-a94f-dd432449c3af tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "df682d97-53c1-4a18-8800-b23f3fc66018" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.787s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.015613] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.911s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.016379] env[61768]: ERROR nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Traceback (most recent call last): [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self.driver.spawn(context, instance, image_meta, [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] vm_ref = self.build_virtual_machine(instance, [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.016379] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] for vif in network_info: [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return self._sync_wrapper(fn, *args, **kwargs) [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self.wait() [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self[:] = self._gt.wait() [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return self._exit_event.wait() [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] result = hub.switch() [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.016728] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return self.greenlet.switch() [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] result = function(*args, **kwargs) [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] return func(*args, **kwargs) [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] raise e [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] nwinfo = self.network_api.allocate_for_instance( [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] created_port_ids = self._update_ports_for_instance( [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] with excutils.save_and_reraise_exception(): [ 524.017061] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] self.force_reraise() [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] raise self.value [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] updated_port = self._update_port( [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] _ensure_no_port_binding_failure(port) [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] raise exception.PortBindingFailed(port_id=port['id']) [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] nova.exception.PortBindingFailed: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. [ 524.017439] env[61768]: ERROR nova.compute.manager [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] [ 524.017725] env[61768]: DEBUG nova.compute.utils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 524.019517] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Build of instance 438d040b-c86f-47ab-8d22-e0e86c41ac43 was re-scheduled: Binding failed for port 5d3aa00a-5c5f-4ca7-a68f-d7f38f1b9bb4, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 524.024223] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 524.024223] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.024223] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquired lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.024223] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 524.025126] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.274s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.030168] env[61768]: INFO nova.compute.claims [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 524.060027] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228529, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070708} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.060027] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 524.060027] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20445100-4b8e-4f5b-b9ec-9f6038343b49 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.082411] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 524.083746] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-75bdfb05-c060-4504-85c1-61b55ffa2526 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.106160] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 524.106160] env[61768]: value = "task-1228530" [ 524.106160] env[61768]: _type = "Task" [ 524.106160] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.114749] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228530, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.129038] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.146126] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Successfully created port: 9888a455-daa3-416d-a640-d0ce412d5c77 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 524.146126] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 524.165035] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.168428] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.168428] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.168428] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.168428] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.168428] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.168858] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.168858] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.168858] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.168858] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.168858] env[61768]: DEBUG nova.virt.hardware [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.169797] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279991d8-6d11-4563-811b-53f282d786a8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.182231] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d092a3f8-0209-482a-ae8c-5d9eeffac18e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.269746] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Acquiring lock "a1b9b618-91b5-429b-8581-9f819e0d81ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.270180] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Lock "a1b9b618-91b5-429b-8581-9f819e0d81ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 524.600763] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.623229] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228530, 'name': ReconfigVM_Task, 'duration_secs': 0.287593} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 524.626024] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Reconfigured VM instance instance-00000008 to attach disk [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 524.626024] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3419cec2-aecb-433c-a367-5d34e93749ca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.638027] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 524.638027] env[61768]: value = "task-1228531" [ 524.638027] env[61768]: _type = "Task" [ 524.638027] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 524.648758] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228531, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 524.668462] env[61768]: ERROR nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. [ 524.668462] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.668462] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.668462] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.668462] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.668462] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.668462] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.668462] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.668462] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.668462] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 524.668462] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.668462] env[61768]: ERROR nova.compute.manager raise self.value [ 524.668462] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.668462] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.668462] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.668462] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.668965] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.668965] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.668965] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. [ 524.668965] env[61768]: ERROR nova.compute.manager [ 524.668965] env[61768]: Traceback (most recent call last): [ 524.668965] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.668965] env[61768]: listener.cb(fileno) [ 524.668965] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.668965] env[61768]: result = function(*args, **kwargs) [ 524.668965] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.668965] env[61768]: return func(*args, **kwargs) [ 524.668965] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.668965] env[61768]: raise e [ 524.668965] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.668965] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 524.668965] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.668965] env[61768]: created_port_ids = self._update_ports_for_instance( [ 524.668965] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.668965] env[61768]: with excutils.save_and_reraise_exception(): [ 524.668965] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.668965] env[61768]: self.force_reraise() [ 524.668965] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.668965] env[61768]: raise self.value [ 524.668965] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.668965] env[61768]: updated_port = self._update_port( [ 524.668965] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.668965] env[61768]: _ensure_no_port_binding_failure(port) [ 524.668965] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.668965] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.669832] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. [ 524.669832] env[61768]: Removing descriptor: 14 [ 524.669832] env[61768]: ERROR nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Traceback (most recent call last): [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] yield resources [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self.driver.spawn(context, instance, image_meta, [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.669832] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] vm_ref = self.build_virtual_machine(instance, [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] for vif in network_info: [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return self._sync_wrapper(fn, *args, **kwargs) [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self.wait() [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self[:] = self._gt.wait() [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return self._exit_event.wait() [ 524.670246] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] result = hub.switch() [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return self.greenlet.switch() [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] result = function(*args, **kwargs) [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return func(*args, **kwargs) [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] raise e [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] nwinfo = self.network_api.allocate_for_instance( [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.670589] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] created_port_ids = self._update_ports_for_instance( [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] with excutils.save_and_reraise_exception(): [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self.force_reraise() [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] raise self.value [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] updated_port = self._update_port( [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] _ensure_no_port_binding_failure(port) [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.670913] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] raise exception.PortBindingFailed(port_id=port['id']) [ 524.671255] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] nova.exception.PortBindingFailed: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. [ 524.671255] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] [ 524.671255] env[61768]: INFO nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Terminating instance [ 524.677349] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Acquiring lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.677349] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Acquired lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.677485] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 524.682817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.880015] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.147339] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228531, 'name': Rename_Task, 'duration_secs': 0.20736} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 525.147796] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 525.148084] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a45d2f94-fcdc-487d-8c02-31a0aec39778 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.156028] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 525.156028] env[61768]: value = "task-1228532" [ 525.156028] env[61768]: _type = "Task" [ 525.156028] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.168149] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.196031] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.196444] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.233441] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.386241] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Releasing lock "refresh_cache-438d040b-c86f-47ab-8d22-e0e86c41ac43" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.386241] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 525.386241] env[61768]: DEBUG nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 525.386241] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 525.389763] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b757f5dd-59aa-43a8-96ca-a59db51d2e25 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.399250] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6acc6b2f-c88c-4357-81ed-36cab16e0df6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.439117] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.441295] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566c86b9-b246-4fd1-b1d2-b1f685fb834c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.453173] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae6d1de-544b-4f14-b968-32c1a0ed0d1f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.469906] env[61768]: DEBUG nova.compute.provider_tree [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 525.487771] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.666512] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228532, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.947742] env[61768]: DEBUG nova.network.neutron [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.974492] env[61768]: DEBUG nova.scheduler.client.report [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 525.990133] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Releasing lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.991762] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 525.992011] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 525.992881] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e333d1c0-6d67-48dd-8622-9712023625a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.003801] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2f604d-2e4b-455b-9b2b-9714f92908f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.029842] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d0e752f-9b39-4d01-bedc-5315927deea8 could not be found. [ 526.030222] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 526.030562] env[61768]: INFO nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 526.030913] env[61768]: DEBUG oslo.service.loopingcall [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.031677] env[61768]: DEBUG nova.compute.manager [-] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.031933] env[61768]: DEBUG nova.network.neutron [-] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 526.068040] env[61768]: DEBUG nova.network.neutron [-] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.123441] env[61768]: DEBUG nova.compute.manager [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Received event network-changed-65e56e3c-761f-4ec3-994c-8b1f396119ed {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.123441] env[61768]: DEBUG nova.compute.manager [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Refreshing instance network info cache due to event network-changed-65e56e3c-761f-4ec3-994c-8b1f396119ed. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 526.123441] env[61768]: DEBUG oslo_concurrency.lockutils [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] Acquiring lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.123441] env[61768]: DEBUG oslo_concurrency.lockutils [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] Acquired lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.123441] env[61768]: DEBUG nova.network.neutron [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Refreshing network info cache for port 65e56e3c-761f-4ec3-994c-8b1f396119ed {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 526.168194] env[61768]: DEBUG oslo_vmware.api [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228532, 'name': PowerOnVM_Task, 'duration_secs': 0.751229} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 526.169164] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 526.169635] env[61768]: INFO nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Took 8.72 seconds to spawn the instance on the hypervisor. [ 526.169978] env[61768]: DEBUG nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 526.172210] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c060744-08a1-4d98-bb50-f7713856082c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.452325] env[61768]: INFO nova.compute.manager [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 438d040b-c86f-47ab-8d22-e0e86c41ac43] Took 1.07 seconds to deallocate network for instance. [ 526.480746] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 526.484022] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 526.485620] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.483s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.569725] env[61768]: DEBUG nova.network.neutron [-] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.697930] env[61768]: INFO nova.compute.manager [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Took 20.99 seconds to build instance. [ 526.820136] env[61768]: DEBUG nova.network.neutron [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.923582] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Acquiring lock "e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.923804] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Lock "e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.992595] env[61768]: DEBUG nova.compute.utils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 526.997750] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 526.997750] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 527.029921] env[61768]: DEBUG nova.network.neutron [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.073045] env[61768]: INFO nova.compute.manager [-] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Took 1.04 seconds to deallocate network for instance. [ 527.075435] env[61768]: DEBUG nova.compute.claims [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 527.076096] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.170936] env[61768]: DEBUG nova.policy [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 527.200676] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb1f143-3be8-4792-9a0c-b8c56b7936cd tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.508s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.276332] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed2fd9e-9e60-4259-b0e9-5f1d43b124e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.285220] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32cfe62c-92a9-4002-83ce-7a16e6e9bd1d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.319993] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea05fd2-7c5a-4ac6-9d7b-4cfd9e1b43dc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.327794] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2c247b5-9d21-4c94-b738-8f5fb58d933d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.343449] env[61768]: DEBUG nova.compute.provider_tree [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.485174] env[61768]: INFO nova.scheduler.client.report [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Deleted allocations for instance 438d040b-c86f-47ab-8d22-e0e86c41ac43 [ 527.503727] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 527.530770] env[61768]: DEBUG oslo_concurrency.lockutils [req-f4157141-1ba4-4941-8fc1-63a77c1f3608 req-25b65374-286f-496d-99b3-67253886ca7e service nova] Releasing lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.647168] env[61768]: ERROR nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. [ 527.647168] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 527.647168] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.647168] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 527.647168] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.647168] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 527.647168] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.647168] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 527.647168] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.647168] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 527.647168] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.647168] env[61768]: ERROR nova.compute.manager raise self.value [ 527.647168] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.647168] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 527.647168] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.647168] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 527.647619] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.647619] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 527.647619] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. [ 527.647619] env[61768]: ERROR nova.compute.manager [ 527.647619] env[61768]: Traceback (most recent call last): [ 527.647619] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 527.647619] env[61768]: listener.cb(fileno) [ 527.647619] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.647619] env[61768]: result = function(*args, **kwargs) [ 527.647861] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.647861] env[61768]: return func(*args, **kwargs) [ 527.647861] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.647861] env[61768]: raise e [ 527.647861] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.647861] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 527.647861] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.647861] env[61768]: created_port_ids = self._update_ports_for_instance( [ 527.647861] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.647861] env[61768]: with excutils.save_and_reraise_exception(): [ 527.647861] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.647861] env[61768]: self.force_reraise() [ 527.647861] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.647861] env[61768]: raise self.value [ 527.647861] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.647861] env[61768]: updated_port = self._update_port( [ 527.647861] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.647861] env[61768]: _ensure_no_port_binding_failure(port) [ 527.647861] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.647861] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 527.647861] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. [ 527.647861] env[61768]: Removing descriptor: 17 [ 527.648510] env[61768]: ERROR nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] Traceback (most recent call last): [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] yield resources [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self.driver.spawn(context, instance, image_meta, [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self._vmops.spawn(context, instance, image_meta, injected_files, [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] vm_ref = self.build_virtual_machine(instance, [ 527.648510] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] vif_infos = vmwarevif.get_vif_info(self._session, [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] for vif in network_info: [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return self._sync_wrapper(fn, *args, **kwargs) [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self.wait() [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self[:] = self._gt.wait() [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return self._exit_event.wait() [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 527.648806] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] result = hub.switch() [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return self.greenlet.switch() [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] result = function(*args, **kwargs) [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return func(*args, **kwargs) [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] raise e [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] nwinfo = self.network_api.allocate_for_instance( [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] created_port_ids = self._update_ports_for_instance( [ 527.649168] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] with excutils.save_and_reraise_exception(): [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self.force_reraise() [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] raise self.value [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] updated_port = self._update_port( [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] _ensure_no_port_binding_failure(port) [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] raise exception.PortBindingFailed(port_id=port['id']) [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] nova.exception.PortBindingFailed: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. [ 527.649599] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] [ 527.649963] env[61768]: INFO nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Terminating instance [ 527.656033] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Acquiring lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.656262] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Acquired lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.656416] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 527.704687] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.847828] env[61768]: DEBUG nova.scheduler.client.report [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 528.000391] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7d8a4efe-8ac5-4be0-9a9f-577b2dfa31f8 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "438d040b-c86f-47ab-8d22-e0e86c41ac43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.124s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.248271] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.352959] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.353894] env[61768]: ERROR nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Traceback (most recent call last): [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self.driver.spawn(context, instance, image_meta, [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self._vmops.spawn(context, instance, image_meta, injected_files, [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] vm_ref = self.build_virtual_machine(instance, [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] vif_infos = vmwarevif.get_vif_info(self._session, [ 528.353894] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] for vif in network_info: [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return self._sync_wrapper(fn, *args, **kwargs) [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self.wait() [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self[:] = self._gt.wait() [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return self._exit_event.wait() [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] result = hub.switch() [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 528.354528] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return self.greenlet.switch() [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] result = function(*args, **kwargs) [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] return func(*args, **kwargs) [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] raise e [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] nwinfo = self.network_api.allocate_for_instance( [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] created_port_ids = self._update_ports_for_instance( [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] with excutils.save_and_reraise_exception(): [ 528.354990] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] self.force_reraise() [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] raise self.value [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] updated_port = self._update_port( [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] _ensure_no_port_binding_failure(port) [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] raise exception.PortBindingFailed(port_id=port['id']) [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] nova.exception.PortBindingFailed: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. [ 528.355475] env[61768]: ERROR nova.compute.manager [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] [ 528.355873] env[61768]: DEBUG nova.compute.utils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 528.359800] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 12.854s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.361862] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Build of instance 0188dbb6-ae91-4601-83a3-3e3bf55fc753 was re-scheduled: Binding failed for port 44ca2d9d-2ce5-47ed-9e5b-8b349b46e524, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 528.362817] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 528.362817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Acquiring lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.362817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Acquired lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.362817] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 528.379326] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Acquiring lock "0103eded-5a0b-4b7f-b98f-63d7866d501c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.379535] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Lock "0103eded-5a0b-4b7f-b98f-63d7866d501c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.447264] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.506190] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.520791] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 528.547257] env[61768]: DEBUG nova.compute.manager [None req-a9110911-6465-4eee-9e4e-f8eafa8c0d57 tempest-ServerDiagnosticsV248Test-1840855018 tempest-ServerDiagnosticsV248Test-1840855018-project-admin] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 528.548995] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ef5a13-6444-40e1-b8ff-8be7b8b748dc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.564914] env[61768]: INFO nova.compute.manager [None req-a9110911-6465-4eee-9e4e-f8eafa8c0d57 tempest-ServerDiagnosticsV248Test-1840855018 tempest-ServerDiagnosticsV248Test-1840855018-project-admin] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Retrieving diagnostics [ 528.569028] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5cbc66b-8d2b-4ab3-9b71-2171802bb9e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.577024] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 528.577024] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 528.577024] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 528.577375] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 528.577375] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 528.577375] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 528.577375] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 528.578899] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 528.578899] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 528.578899] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 528.579077] env[61768]: DEBUG nova.virt.hardware [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 528.580254] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d877e1-ae07-4c5a-88d4-2f86ff221a71 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.620110] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bf92d01-9b88-4285-ad54-2e88c6d28b43 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.728455] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "70e6d944-f978-4d5b-b396-e0e913db0485" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.728455] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "70e6d944-f978-4d5b-b396-e0e913db0485" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.736186] env[61768]: INFO nova.compute.manager [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Rebuilding instance [ 528.756999] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Successfully created port: 0591a789-7773-484e-9663-e763b0bb4667 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 528.787797] env[61768]: DEBUG nova.compute.manager [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 528.788682] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caeab6c5-9215-4e0e-ac1c-2888735542ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.846636] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.912624] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.029000] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.162982] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.305394] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 529.306060] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3aeb41dd-1697-4bc8-994c-2cea7c018a1a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.316398] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 529.316398] env[61768]: value = "task-1228533" [ 529.316398] env[61768]: _type = "Task" [ 529.316398] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.329604] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228533, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.350675] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Releasing lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.351131] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 529.351333] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 529.351630] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-164f4d5d-e1cc-4e16-90da-6315818bcc51 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.365255] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa815d9-6b72-4799-8da5-04195e4a864c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.406991] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1b694ac3-385b-4083-a35c-542853354202 could not be found. [ 529.410319] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 529.410319] env[61768]: INFO nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Took 0.06 seconds to destroy the instance on the hypervisor. [ 529.410319] env[61768]: DEBUG oslo.service.loopingcall [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.410319] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance ca465e62-c190-4d67-9dec-87ea1aa8f194 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 529.410319] env[61768]: DEBUG nova.compute.manager [-] [instance: 1b694ac3-385b-4083-a35c-542853354202] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.411096] env[61768]: DEBUG nova.network.neutron [-] [instance: 1b694ac3-385b-4083-a35c-542853354202] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 529.446837] env[61768]: DEBUG nova.network.neutron [-] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.603109] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "fcddfd64-74bf-4a05-9799-0b7fee7416b2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.603109] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "fcddfd64-74bf-4a05-9799-0b7fee7416b2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.666165] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Releasing lock "refresh_cache-0188dbb6-ae91-4601-83a3-3e3bf55fc753" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 529.667275] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 529.667502] env[61768]: DEBUG nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 529.668270] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 529.717160] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 529.833168] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228533, 'name': PowerOffVM_Task, 'duration_secs': 0.125172} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 529.833429] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 529.833688] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 529.834831] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e759e2c-4454-4450-a3aa-83cf3dbdc905 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.846849] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 529.847222] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc511575-9cf6-49b1-a487-30534775d8b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.873491] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 529.873741] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 529.873950] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleting the datastore file [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 529.874257] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0fe3cdaf-42d0-4442-a3a3-0d155619205a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.887787] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 529.887787] env[61768]: value = "task-1228535" [ 529.887787] env[61768]: _type = "Task" [ 529.887787] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.899675] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228535, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.917694] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 0188dbb6-ae91-4601-83a3-3e3bf55fc753 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 529.917844] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance c3b90a4d-e885-453c-9f8e-799adb2400a0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 529.917965] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 0408fe1e-e4fa-4649-83dc-afdd335ee30d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 529.918114] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7d0e752f-9b39-4d01-bedc-5315927deea8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 529.918234] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 1b694ac3-385b-4083-a35c-542853354202 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 529.918375] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 417480ad-61d9-447d-a84d-0cf5df94f4bc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 529.949921] env[61768]: DEBUG nova.network.neutron [-] [instance: 1b694ac3-385b-4083-a35c-542853354202] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 529.997956] env[61768]: DEBUG nova.compute.manager [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Received event network-vif-deleted-65e56e3c-761f-4ec3-994c-8b1f396119ed {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.998295] env[61768]: DEBUG nova.compute.manager [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] [instance: 1b694ac3-385b-4083-a35c-542853354202] Received event network-changed-9888a455-daa3-416d-a640-d0ce412d5c77 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.998433] env[61768]: DEBUG nova.compute.manager [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] [instance: 1b694ac3-385b-4083-a35c-542853354202] Refreshing instance network info cache due to event network-changed-9888a455-daa3-416d-a640-d0ce412d5c77. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 529.998986] env[61768]: DEBUG oslo_concurrency.lockutils [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] Acquiring lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.999508] env[61768]: DEBUG oslo_concurrency.lockutils [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] Acquired lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.999714] env[61768]: DEBUG nova.network.neutron [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] [instance: 1b694ac3-385b-4083-a35c-542853354202] Refreshing network info cache for port 9888a455-daa3-416d-a640-d0ce412d5c77 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 530.219826] env[61768]: DEBUG nova.network.neutron [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.400951] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228535, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.108584} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.400951] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 530.401714] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 530.401714] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 530.424595] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance eb8203da-3427-48a8-9ec0-a9d12d8c44d2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 530.453045] env[61768]: INFO nova.compute.manager [-] [instance: 1b694ac3-385b-4083-a35c-542853354202] Took 1.04 seconds to deallocate network for instance. [ 530.455680] env[61768]: DEBUG nova.compute.claims [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 530.455827] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.545138] env[61768]: DEBUG nova.network.neutron [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 530.722998] env[61768]: INFO nova.compute.manager [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] [instance: 0188dbb6-ae91-4601-83a3-3e3bf55fc753] Took 1.06 seconds to deallocate network for instance. [ 530.779720] env[61768]: DEBUG nova.network.neutron [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] [instance: 1b694ac3-385b-4083-a35c-542853354202] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.929523] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7bc2075c-2cb2-41c2-9681-780689c1d080 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 531.135540] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "ca465e62-c190-4d67-9dec-87ea1aa8f194" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.136216] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "ca465e62-c190-4d67-9dec-87ea1aa8f194" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.136216] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "ca465e62-c190-4d67-9dec-87ea1aa8f194-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.136216] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "ca465e62-c190-4d67-9dec-87ea1aa8f194-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.136451] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "ca465e62-c190-4d67-9dec-87ea1aa8f194-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.139589] env[61768]: INFO nova.compute.manager [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Terminating instance [ 531.142660] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "refresh_cache-ca465e62-c190-4d67-9dec-87ea1aa8f194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 531.142800] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquired lock "refresh_cache-ca465e62-c190-4d67-9dec-87ea1aa8f194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 531.142968] env[61768]: DEBUG nova.network.neutron [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 531.181800] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Acquiring lock "f0be276a-ddad-4ae0-a232-d0fb8463a3bc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.182035] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Lock "f0be276a-ddad-4ae0-a232-d0fb8463a3bc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.284138] env[61768]: DEBUG oslo_concurrency.lockutils [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] Releasing lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 531.284138] env[61768]: DEBUG nova.compute.manager [req-1975352c-c14d-436c-b524-b38fb85db054 req-72a87457-f013-46d8-a66b-9b14869a4872 service nova] [instance: 1b694ac3-385b-4083-a35c-542853354202] Received event network-vif-deleted-9888a455-daa3-416d-a640-d0ce412d5c77 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 531.432982] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance ca31b2ec-0bfb-47d9-9f35-c54659fa7799 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 531.462428] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.465380] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.465380] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.465380] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.465380] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.465380] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.465551] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.465551] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.465551] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.465551] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.465551] env[61768]: DEBUG nova.virt.hardware [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.466468] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e664c2a-9292-4303-af55-0355429ab92d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.477363] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b6f8781-44e8-4dc2-bcd4-48b70cfee567 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.494240] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 531.500387] env[61768]: DEBUG oslo.service.loopingcall [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 531.500480] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 531.500615] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9213f2ac-a0b7-4c6f-87f5-d43f6e88bfe2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.518509] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 531.518509] env[61768]: value = "task-1228536" [ 531.518509] env[61768]: _type = "Task" [ 531.518509] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.526730] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228536, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.553578] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "aadc35e7-7aea-4f86-be3a-9b5322f350a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.553778] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "aadc35e7-7aea-4f86-be3a-9b5322f350a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.729018] env[61768]: DEBUG nova.network.neutron [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 531.780745] env[61768]: INFO nova.scheduler.client.report [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Deleted allocations for instance 0188dbb6-ae91-4601-83a3-3e3bf55fc753 [ 531.937702] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 531.963624] env[61768]: DEBUG nova.network.neutron [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 532.037232] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228536, 'name': CreateVM_Task, 'duration_secs': 0.286547} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.037419] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 532.037823] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.037977] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.038342] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 532.038647] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c05a5c7e-f991-4243-8b3c-ce92fcaca3a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.049660] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 532.049660] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b01ed5-5bd3-32ae-e2f4-53c30c840781" [ 532.049660] env[61768]: _type = "Task" [ 532.049660] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.058054] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b01ed5-5bd3-32ae-e2f4-53c30c840781, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.289122] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a668a793-5ee2-4781-a833-33883d6b9e86 tempest-FloatingIPsAssociationTestJSON-2122556692 tempest-FloatingIPsAssociationTestJSON-2122556692-project-member] Lock "0188dbb6-ae91-4601-83a3-3e3bf55fc753" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.763s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.441931] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 1331a447-7be8-42b8-9d59-6b97742e49d1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 532.465662] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Releasing lock "refresh_cache-ca465e62-c190-4d67-9dec-87ea1aa8f194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.466114] env[61768]: DEBUG nova.compute.manager [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 532.467579] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 532.467579] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f4eb425-188a-4ff4-a617-6f654adcb21e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.476830] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 532.477117] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-86092e1e-f568-431e-9960-d67ae6799c94 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.484868] env[61768]: DEBUG oslo_vmware.api [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 532.484868] env[61768]: value = "task-1228537" [ 532.484868] env[61768]: _type = "Task" [ 532.484868] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.497284] env[61768]: ERROR nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. [ 532.497284] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.497284] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.497284] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.497284] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.497284] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.497284] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.497284] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.497284] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.497284] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 532.497284] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.497284] env[61768]: ERROR nova.compute.manager raise self.value [ 532.497284] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.497284] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.497284] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.497284] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.498138] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.498138] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.498138] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. [ 532.498138] env[61768]: ERROR nova.compute.manager [ 532.498138] env[61768]: Traceback (most recent call last): [ 532.498138] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.498138] env[61768]: listener.cb(fileno) [ 532.498138] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.498138] env[61768]: result = function(*args, **kwargs) [ 532.498138] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.498138] env[61768]: return func(*args, **kwargs) [ 532.498138] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.498138] env[61768]: raise e [ 532.498138] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.498138] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 532.498138] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.498138] env[61768]: created_port_ids = self._update_ports_for_instance( [ 532.498138] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.498138] env[61768]: with excutils.save_and_reraise_exception(): [ 532.498138] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.498138] env[61768]: self.force_reraise() [ 532.498138] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.498138] env[61768]: raise self.value [ 532.498138] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.498138] env[61768]: updated_port = self._update_port( [ 532.498138] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.498138] env[61768]: _ensure_no_port_binding_failure(port) [ 532.498138] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.498138] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.499456] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. [ 532.499456] env[61768]: Removing descriptor: 14 [ 532.499456] env[61768]: ERROR nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Traceback (most recent call last): [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] yield resources [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self.driver.spawn(context, instance, image_meta, [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.499456] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] vm_ref = self.build_virtual_machine(instance, [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] for vif in network_info: [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return self._sync_wrapper(fn, *args, **kwargs) [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self.wait() [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self[:] = self._gt.wait() [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return self._exit_event.wait() [ 532.500170] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] result = hub.switch() [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return self.greenlet.switch() [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] result = function(*args, **kwargs) [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return func(*args, **kwargs) [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] raise e [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] nwinfo = self.network_api.allocate_for_instance( [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.500790] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] created_port_ids = self._update_ports_for_instance( [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] with excutils.save_and_reraise_exception(): [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self.force_reraise() [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] raise self.value [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] updated_port = self._update_port( [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] _ensure_no_port_binding_failure(port) [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.501395] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] raise exception.PortBindingFailed(port_id=port['id']) [ 532.501913] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] nova.exception.PortBindingFailed: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. [ 532.501913] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] [ 532.501913] env[61768]: INFO nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Terminating instance [ 532.504698] env[61768]: DEBUG oslo_vmware.api [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228537, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.505512] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.505664] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.505828] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 532.563079] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b01ed5-5bd3-32ae-e2f4-53c30c840781, 'name': SearchDatastore_Task, 'duration_secs': 0.009681} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.563453] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 532.563744] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 532.563980] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.564157] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.564586] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 532.564853] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2754db88-ef0f-4340-a7be-fac646d7b7fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.574107] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 532.575525] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 532.577846] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1d3d13b-3095-43f6-8462-2c6aace37a83 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.584714] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Acquiring lock "21099fc3-dd66-4871-a87c-e0d2c56a7818" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.584714] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Lock "21099fc3-dd66-4871-a87c-e0d2c56a7818" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.589749] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 532.589749] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a71214-af41-d92d-746b-68fd31ff8dc3" [ 532.589749] env[61768]: _type = "Task" [ 532.589749] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.598477] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a71214-af41-d92d-746b-68fd31ff8dc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.710771] env[61768]: DEBUG nova.compute.manager [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Received event network-changed-0591a789-7773-484e-9663-e763b0bb4667 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 532.710914] env[61768]: DEBUG nova.compute.manager [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Refreshing instance network info cache due to event network-changed-0591a789-7773-484e-9663-e763b0bb4667. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 532.711107] env[61768]: DEBUG oslo_concurrency.lockutils [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] Acquiring lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.795055] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 532.948997] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance a1b9b618-91b5-429b-8581-9f819e0d81ee has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 532.995984] env[61768]: DEBUG oslo_vmware.api [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228537, 'name': PowerOffVM_Task, 'duration_secs': 0.226054} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.996521] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 532.996740] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 532.997045] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-97653cdf-7194-489e-8f8e-03c266e3c939 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.027531] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 533.027531] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 533.027531] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Deleting the datastore file [datastore2] ca465e62-c190-4d67-9dec-87ea1aa8f194 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 533.027531] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd0f5645-48a2-425d-948e-275d0fa422b7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.036178] env[61768]: DEBUG oslo_vmware.api [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for the task: (returnval){ [ 533.036178] env[61768]: value = "task-1228539" [ 533.036178] env[61768]: _type = "Task" [ 533.036178] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.044200] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.056316] env[61768]: DEBUG oslo_vmware.api [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228539, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.106633] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a71214-af41-d92d-746b-68fd31ff8dc3, 'name': SearchDatastore_Task, 'duration_secs': 0.00917} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.106633] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-632b41e8-2e6f-480f-9d29-8f06a716b5e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.111867] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 533.111867] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f0b7d5-ca9c-edce-39c9-ad78dd0185dd" [ 533.111867] env[61768]: _type = "Task" [ 533.111867] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.122455] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f0b7d5-ca9c-edce-39c9-ad78dd0185dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.324319] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.332133] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.453920] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 533.550968] env[61768]: DEBUG oslo_vmware.api [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Task: {'id': task-1228539, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106076} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.551218] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 533.551398] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 533.551563] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 533.551730] env[61768]: INFO nova.compute.manager [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Took 1.09 seconds to destroy the instance on the hypervisor. [ 533.551971] env[61768]: DEBUG oslo.service.loopingcall [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.552186] env[61768]: DEBUG nova.compute.manager [-] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.552283] env[61768]: DEBUG nova.network.neutron [-] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 533.592637] env[61768]: DEBUG nova.network.neutron [-] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.623194] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f0b7d5-ca9c-edce-39c9-ad78dd0185dd, 'name': SearchDatastore_Task, 'duration_secs': 0.009946} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.623460] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.623711] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 533.623968] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6eec4256-4e62-4b16-9d3e-c77ae05e7695 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.633494] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 533.633494] env[61768]: value = "task-1228540" [ 533.633494] env[61768]: _type = "Task" [ 533.633494] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.644064] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.836325] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.836325] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 533.836325] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 533.836325] env[61768]: DEBUG oslo_concurrency.lockutils [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] Acquired lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.836325] env[61768]: DEBUG nova.network.neutron [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Refreshing network info cache for port 0591a789-7773-484e-9663-e763b0bb4667 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 533.837511] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16b68f93-30a6-47b4-9dc8-367fe085deff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.852825] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8ef3418-db09-45c1-9aee-206a214e3ef4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.881967] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 417480ad-61d9-447d-a84d-0cf5df94f4bc could not be found. [ 533.882300] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 533.882728] env[61768]: INFO nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 533.883147] env[61768]: DEBUG oslo.service.loopingcall [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.883364] env[61768]: DEBUG nova.compute.manager [-] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.883502] env[61768]: DEBUG nova.network.neutron [-] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 533.932344] env[61768]: DEBUG nova.network.neutron [-] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.958309] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance e4a7dc76-99a6-4cc2-97c8-f86d72d00d12 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.100184] env[61768]: DEBUG nova.network.neutron [-] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.146741] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474385} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.147260] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 534.147491] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 534.147963] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a2020b1e-6810-420d-818d-cfdcb43986e8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.157880] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 534.157880] env[61768]: value = "task-1228541" [ 534.157880] env[61768]: _type = "Task" [ 534.157880] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.168296] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.396634] env[61768]: DEBUG nova.network.neutron [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.438218] env[61768]: DEBUG nova.network.neutron [-] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.460887] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 0103eded-5a0b-4b7f-b98f-63d7866d501c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 534.604112] env[61768]: INFO nova.compute.manager [-] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Took 1.05 seconds to deallocate network for instance. [ 534.667676] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095379} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.667676] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 534.668443] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4650cb7-a3aa-4e0c-9c6e-2f9e08d480ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.699499] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 534.700526] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3635801-b87d-423d-9d9c-72386e250ece {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.723822] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 534.723822] env[61768]: value = "task-1228542" [ 534.723822] env[61768]: _type = "Task" [ 534.723822] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.733021] env[61768]: DEBUG nova.network.neutron [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.739070] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228542, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.946577] env[61768]: INFO nova.compute.manager [-] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Took 1.06 seconds to deallocate network for instance. [ 534.948082] env[61768]: DEBUG nova.compute.claims [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 534.948297] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.969332] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 70e6d944-f978-4d5b-b396-e0e913db0485 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.117772] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.121550] env[61768]: DEBUG nova.compute.manager [req-91ee5c86-c966-479f-8571-676acee82914 req-fea1eb16-86eb-4135-b3a9-c85dd6066553 service nova] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Received event network-vif-deleted-0591a789-7773-484e-9663-e763b0bb4667 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 535.234305] env[61768]: DEBUG oslo_concurrency.lockutils [req-c56bb564-19e8-41ad-b97d-10c604dcf720 req-aa977c09-5f72-472c-ae6c-dfe6ae77863f service nova] Releasing lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.234305] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228542, 'name': ReconfigVM_Task, 'duration_secs': 0.314192} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.234305] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Reconfigured VM instance instance-00000008 to attach disk [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 535.234899] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-326a827f-dea4-489c-8214-f6f56ff1f1d6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.241478] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 535.241478] env[61768]: value = "task-1228543" [ 535.241478] env[61768]: _type = "Task" [ 535.241478] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.250633] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228543, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.397844] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "83afb27a-7984-4982-9880-f5b0c6553787" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.398107] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "83afb27a-7984-4982-9880-f5b0c6553787" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.475027] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance fcddfd64-74bf-4a05-9799-0b7fee7416b2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 535.475027] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 535.475027] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 535.754134] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228543, 'name': Rename_Task, 'duration_secs': 0.137711} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.757285] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 535.757860] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b697cc17-2075-4136-88f0-53698ab2666c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.765584] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 535.765584] env[61768]: value = "task-1228544" [ 535.765584] env[61768]: _type = "Task" [ 535.765584] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.777960] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228544, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.825134] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bb138d6-6ef5-4272-bb12-0bebc52d71a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.833009] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1e022b-25a6-4b49-a0d5-fd98ccd8b55e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.868100] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195b9f08-a347-435a-9912-92c871989cba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.878647] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b42460-9155-48ff-9ea9-d95d74ab89db {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.890680] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.280918] env[61768]: DEBUG oslo_vmware.api [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228544, 'name': PowerOnVM_Task, 'duration_secs': 0.440838} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.281595] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 536.281595] env[61768]: DEBUG nova.compute.manager [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 536.286042] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2972f36-9950-45ee-8823-a147a0bc3f74 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.395396] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.808643] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.902977] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 536.903261] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.544s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.903637] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.269s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.905296] env[61768]: INFO nova.compute.claims [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.163232] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Acquiring lock "8ba075db-5095-4c38-afbc-85e9d35d6194" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.163764] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Lock "8ba075db-5095-4c38-afbc-85e9d35d6194" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.464688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Acquiring lock "e6922bb7-c2df-468a-aa5b-d03e274b8ecc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.464920] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Lock "e6922bb7-c2df-468a-aa5b-d03e274b8ecc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.414241] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e47b48a6-add8-42e5-b6ca-eaac3a6bc83e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.424058] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f477a4-4d81-41e4-89c1-b352fe32084d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.462256] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764562a9-4673-47fc-91f6-e77b380e2eb0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.471691] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd868a17-5003-41c4-a21f-67c78b25c9da {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.487763] env[61768]: DEBUG nova.compute.provider_tree [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.957813] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Acquiring lock "07d8d6b6-3cc5-48da-9d4d-563a5fe27442" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.957813] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Lock "07d8d6b6-3cc5-48da-9d4d-563a5fe27442" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.991100] env[61768]: DEBUG nova.scheduler.client.report [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.340675] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "f81aabf6-ae46-405c-b101-12ca707a0567" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.342091] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "f81aabf6-ae46-405c-b101-12ca707a0567" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.424023] env[61768]: INFO nova.compute.manager [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Rebuilding instance [ 539.487902] env[61768]: DEBUG nova.compute.manager [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 539.488849] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5802639-ba08-400a-b39a-b35763b9e9db {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.497299] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.594s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.497862] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.504492] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.835s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.505231] env[61768]: INFO nova.compute.claims [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.005637] env[61768]: DEBUG nova.compute.utils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.007617] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 540.008173] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 540.010633] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 540.013318] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-af2b678a-eee5-4f9e-a24a-52b0c333eb24 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.021983] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 540.021983] env[61768]: value = "task-1228545" [ 540.021983] env[61768]: _type = "Task" [ 540.021983] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.032986] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228545, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.111982] env[61768]: DEBUG nova.policy [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2eb3cbf61bc4bdbaaaf29cdf78e4853', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76967cef64d8478ab599e0f336d67535', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.497085] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Acquiring lock "4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.497418] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Lock "4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.509328] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.533454] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228545, 'name': PowerOffVM_Task, 'duration_secs': 0.157821} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.533723] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 540.533932] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 540.535094] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2c6537-2d8e-4d77-a6c6-ff391360f784 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.543846] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 540.544804] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4679107-5298-44b6-bd35-968468fab08e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.573612] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 540.573612] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 540.573612] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Deleting the datastore file [datastore1] 0408fe1e-e4fa-4649-83dc-afdd335ee30d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 540.573612] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-948dc0cb-7f72-4030-9198-105d88e32b6d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.580612] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 540.580612] env[61768]: value = "task-1228547" [ 540.580612] env[61768]: _type = "Task" [ 540.580612] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.591994] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228547, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.814638] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Successfully created port: 47866dda-3b73-4baa-8653-848543575342 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.957168] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-137fc863-32cd-4706-a85f-ea72c752b2ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.964312] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ef8450-7b0d-41fd-a08a-40c2aa416641 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.995955] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8bcb75-f686-4a19-b2f3-d52c39e54f1d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.003906] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d753203e-8b4a-45cc-8391-8d94ca69673d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.021543] env[61768]: DEBUG nova.compute.provider_tree [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.090654] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228547, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185368} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.090915] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 541.091110] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 541.091289] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 541.165759] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "7d35cf35-d264-468d-a9ab-7b41dd020023" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.166126] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "7d35cf35-d264-468d-a9ab-7b41dd020023" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.415509] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "1eb76ef5-377c-4447-92da-4c61345c0070" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.415695] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "1eb76ef5-377c-4447-92da-4c61345c0070" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.525200] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.528735] env[61768]: DEBUG nova.scheduler.client.report [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.562979] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.562979] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.562979] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.563359] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.563660] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.563944] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.564533] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.564813] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.565176] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.567025] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.567025] env[61768]: DEBUG nova.virt.hardware [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.567989] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fdc08f-eea6-466a-9b33-0738da770099 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.576165] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76395da4-0ad4-4182-be2c-620d193045d2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.035900] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.532s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.036455] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 542.041219] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.525s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.143067] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.143409] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.143609] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.143819] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.146941] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.146941] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.146941] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.146941] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.146941] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.147186] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.147186] env[61768]: DEBUG nova.virt.hardware [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.147186] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20bc6d5e-defe-42a0-ba96-6d862092662e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.157404] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf67080a-4b54-48fd-835c-369b61efdeb1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.172095] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 542.178601] env[61768]: DEBUG oslo.service.loopingcall [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.179232] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 542.179453] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd094e92-3237-472d-aaed-99f11a6e64c3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.197168] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "6035033d-9e22-49bd-a42c-3faeeb7c5d21" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.197411] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "6035033d-9e22-49bd-a42c-3faeeb7c5d21" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.201675] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 542.201675] env[61768]: value = "task-1228548" [ 542.201675] env[61768]: _type = "Task" [ 542.201675] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.209819] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228548, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.545924] env[61768]: DEBUG nova.compute.utils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.547977] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.549217] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 542.612827] env[61768]: DEBUG nova.policy [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7f7389d526c4ce9bc8ecf062b9921e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '180da0bcd5954eb2bc89a20cf8f42bb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.720616] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228548, 'name': CreateVM_Task, 'duration_secs': 0.316315} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.720898] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 542.721495] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.721690] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.722566] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 542.723037] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f9d5c45-6d43-4b37-895b-e781bd4f667c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.728767] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 542.728767] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f9cc33-e968-8f59-549e-6c51802402a1" [ 542.728767] env[61768]: _type = "Task" [ 542.728767] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.741555] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f9cc33-e968-8f59-549e-6c51802402a1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.876895] env[61768]: ERROR nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. [ 542.876895] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 542.876895] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.876895] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 542.876895] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.876895] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 542.876895] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.876895] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 542.876895] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.876895] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 542.876895] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.876895] env[61768]: ERROR nova.compute.manager raise self.value [ 542.876895] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.876895] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 542.876895] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.876895] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 542.877425] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.877425] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 542.877425] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. [ 542.877425] env[61768]: ERROR nova.compute.manager [ 542.877425] env[61768]: Traceback (most recent call last): [ 542.877425] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 542.877425] env[61768]: listener.cb(fileno) [ 542.877425] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.877425] env[61768]: result = function(*args, **kwargs) [ 542.877425] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.877425] env[61768]: return func(*args, **kwargs) [ 542.877425] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.877425] env[61768]: raise e [ 542.877425] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.877425] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 542.877425] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.877425] env[61768]: created_port_ids = self._update_ports_for_instance( [ 542.877425] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.877425] env[61768]: with excutils.save_and_reraise_exception(): [ 542.877425] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.877425] env[61768]: self.force_reraise() [ 542.877425] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.877425] env[61768]: raise self.value [ 542.877425] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.877425] env[61768]: updated_port = self._update_port( [ 542.877425] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.877425] env[61768]: _ensure_no_port_binding_failure(port) [ 542.877425] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.877425] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 542.878578] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. [ 542.878578] env[61768]: Removing descriptor: 14 [ 542.879849] env[61768]: ERROR nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Traceback (most recent call last): [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] yield resources [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self.driver.spawn(context, instance, image_meta, [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] vm_ref = self.build_virtual_machine(instance, [ 542.879849] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] for vif in network_info: [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return self._sync_wrapper(fn, *args, **kwargs) [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self.wait() [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self[:] = self._gt.wait() [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return self._exit_event.wait() [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.880352] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] result = hub.switch() [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return self.greenlet.switch() [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] result = function(*args, **kwargs) [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return func(*args, **kwargs) [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] raise e [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] nwinfo = self.network_api.allocate_for_instance( [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] created_port_ids = self._update_ports_for_instance( [ 542.880719] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] with excutils.save_and_reraise_exception(): [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self.force_reraise() [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] raise self.value [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] updated_port = self._update_port( [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] _ensure_no_port_binding_failure(port) [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] raise exception.PortBindingFailed(port_id=port['id']) [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] nova.exception.PortBindingFailed: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. [ 542.881112] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] [ 542.881481] env[61768]: INFO nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Terminating instance [ 542.883855] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.884038] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquired lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.884205] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 542.907118] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "1def011b-674f-4336-967c-96d26d48aa6c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.910194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "1def011b-674f-4336-967c-96d26d48aa6c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.058458] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 543.109514] env[61768]: DEBUG nova.compute.manager [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Received event network-changed-47866dda-3b73-4baa-8653-848543575342 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 543.109984] env[61768]: DEBUG nova.compute.manager [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Refreshing instance network info cache due to event network-changed-47866dda-3b73-4baa-8653-848543575342. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.109984] env[61768]: DEBUG oslo_concurrency.lockutils [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] Acquiring lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.110808] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Successfully created port: 889ff1e9-de2a-434c-9ab7-9a15c77a296b {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 543.122019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dddc14a-e9dd-4851-9db8-bc448b79d0e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.130756] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8d4f0f-c477-41fd-9951-c61bbaaf0a73 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.166773] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ced0e7-d6e8-4d6e-99fc-7470d7076be2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.175663] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076d69e3-01c8-40ae-9760-c11fde926977 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.191224] env[61768]: DEBUG nova.compute.provider_tree [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.240323] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f9cc33-e968-8f59-549e-6c51802402a1, 'name': SearchDatastore_Task, 'duration_secs': 0.009976} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.240323] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.240323] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 543.240765] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.240765] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.240882] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 543.241116] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b59e744-4459-4bf4-86af-59257d14da5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.248801] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 543.249029] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 543.249893] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eb433027-6856-4083-b9bf-bb4e4dd1e8d6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.255036] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 543.255036] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]521fb09e-a342-7782-430f-dbfc419729d5" [ 543.255036] env[61768]: _type = "Task" [ 543.255036] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.265270] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]521fb09e-a342-7782-430f-dbfc419729d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.414962] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.539611] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.694064] env[61768]: DEBUG nova.scheduler.client.report [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.767326] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]521fb09e-a342-7782-430f-dbfc419729d5, 'name': SearchDatastore_Task, 'duration_secs': 0.008151} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.768258] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb34ba05-2dc0-47db-a4c4-0128fb76bf7a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.773699] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 543.773699] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d54e5b-7d15-18bc-153e-20cc91088055" [ 543.773699] env[61768]: _type = "Task" [ 543.773699] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.785058] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d54e5b-7d15-18bc-153e-20cc91088055, 'name': SearchDatastore_Task, 'duration_secs': 0.008713} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.785708] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.785708] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 543.786171] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05073fe1-b420-4633-b0ed-a676557cb12c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.792938] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 543.792938] env[61768]: value = "task-1228549" [ 543.792938] env[61768]: _type = "Task" [ 543.792938] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.801092] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228549, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.824589] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "496c3799-7350-4258-a456-58bd3ba9b4d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.824589] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "496c3799-7350-4258-a456-58bd3ba9b4d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.042321] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Releasing lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.042793] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.043017] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 544.043390] env[61768]: DEBUG oslo_concurrency.lockutils [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] Acquired lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.043802] env[61768]: DEBUG nova.network.neutron [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Refreshing network info cache for port 47866dda-3b73-4baa-8653-848543575342 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 544.045425] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-18eddd9b-827b-4abf-9d41-14259a33d70b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.057520] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f59689-50d7-433b-ba08-98f3c2868324 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.076725] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 544.096359] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eb8203da-3427-48a8-9ec0-a9d12d8c44d2 could not be found. [ 544.096359] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 544.096359] env[61768]: INFO nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Took 0.05 seconds to destroy the instance on the hypervisor. [ 544.096359] env[61768]: DEBUG oslo.service.loopingcall [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.099261] env[61768]: DEBUG nova.compute.manager [-] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.099397] env[61768]: DEBUG nova.network.neutron [-] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 544.125831] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.126353] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.126656] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.126879] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.127044] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.127199] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.127566] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.127566] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.127727] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.127885] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.129250] env[61768]: DEBUG nova.virt.hardware [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.130336] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df32f2c-13c9-4ce7-a126-df5038f63aff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.142198] env[61768]: DEBUG nova.network.neutron [-] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.142971] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81f2901-915e-4df0-aa47-44cf088ddb01 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.199727] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.161s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.201514] env[61768]: ERROR nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Traceback (most recent call last): [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self.driver.spawn(context, instance, image_meta, [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] vm_ref = self.build_virtual_machine(instance, [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.201514] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] for vif in network_info: [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return self._sync_wrapper(fn, *args, **kwargs) [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self.wait() [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self[:] = self._gt.wait() [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return self._exit_event.wait() [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] result = hub.switch() [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.201914] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return self.greenlet.switch() [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] result = function(*args, **kwargs) [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] return func(*args, **kwargs) [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] raise e [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] nwinfo = self.network_api.allocate_for_instance( [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] created_port_ids = self._update_ports_for_instance( [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] with excutils.save_and_reraise_exception(): [ 544.202389] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] self.force_reraise() [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] raise self.value [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] updated_port = self._update_port( [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] _ensure_no_port_binding_failure(port) [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] raise exception.PortBindingFailed(port_id=port['id']) [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] nova.exception.PortBindingFailed: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. [ 544.202774] env[61768]: ERROR nova.compute.manager [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] [ 544.203136] env[61768]: DEBUG nova.compute.utils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.203136] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.520s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.204336] env[61768]: INFO nova.compute.claims [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.207512] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Build of instance c3b90a4d-e885-453c-9f8e-799adb2400a0 was re-scheduled: Binding failed for port 9fbb9c99-fa97-43b0-a4b0-6774ff75b899, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 544.207948] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 544.208073] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.208225] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquired lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.208386] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.303125] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228549, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445893} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.306363] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 544.306363] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 544.306363] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dcdfecd9-9d8e-4979-83fd-c124bb9ed429 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.310240] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 544.310240] env[61768]: value = "task-1228550" [ 544.310240] env[61768]: _type = "Task" [ 544.310240] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.318887] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228550, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.548126] env[61768]: ERROR nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. [ 544.548126] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 544.548126] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.548126] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 544.548126] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.548126] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 544.548126] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.548126] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 544.548126] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.548126] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 544.548126] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.548126] env[61768]: ERROR nova.compute.manager raise self.value [ 544.548126] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.548126] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 544.548126] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.548126] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 544.548919] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.548919] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 544.548919] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. [ 544.548919] env[61768]: ERROR nova.compute.manager [ 544.548919] env[61768]: Traceback (most recent call last): [ 544.548919] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 544.548919] env[61768]: listener.cb(fileno) [ 544.548919] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.548919] env[61768]: result = function(*args, **kwargs) [ 544.548919] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.548919] env[61768]: return func(*args, **kwargs) [ 544.548919] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.548919] env[61768]: raise e [ 544.548919] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.548919] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 544.548919] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.548919] env[61768]: created_port_ids = self._update_ports_for_instance( [ 544.548919] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.548919] env[61768]: with excutils.save_and_reraise_exception(): [ 544.548919] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.548919] env[61768]: self.force_reraise() [ 544.548919] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.548919] env[61768]: raise self.value [ 544.548919] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.548919] env[61768]: updated_port = self._update_port( [ 544.548919] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.548919] env[61768]: _ensure_no_port_binding_failure(port) [ 544.548919] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.548919] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 544.550389] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. [ 544.550389] env[61768]: Removing descriptor: 17 [ 544.552465] env[61768]: ERROR nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Traceback (most recent call last): [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] yield resources [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self.driver.spawn(context, instance, image_meta, [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] vm_ref = self.build_virtual_machine(instance, [ 544.552465] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] for vif in network_info: [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return self._sync_wrapper(fn, *args, **kwargs) [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self.wait() [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self[:] = self._gt.wait() [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return self._exit_event.wait() [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.552848] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] result = hub.switch() [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return self.greenlet.switch() [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] result = function(*args, **kwargs) [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return func(*args, **kwargs) [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] raise e [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] nwinfo = self.network_api.allocate_for_instance( [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] created_port_ids = self._update_ports_for_instance( [ 544.553293] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] with excutils.save_and_reraise_exception(): [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self.force_reraise() [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] raise self.value [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] updated_port = self._update_port( [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] _ensure_no_port_binding_failure(port) [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] raise exception.PortBindingFailed(port_id=port['id']) [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] nova.exception.PortBindingFailed: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. [ 544.554194] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] [ 544.554596] env[61768]: INFO nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Terminating instance [ 544.556419] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.556583] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.556748] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.567335] env[61768]: DEBUG nova.network.neutron [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.648320] env[61768]: DEBUG nova.network.neutron [-] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.689555] env[61768]: DEBUG nova.network.neutron [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.734767] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.820183] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228550, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063942} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.821018] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 544.822659] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0f86ed3-3528-4e09-bd39-51af8dc65a72 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.850022] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Reconfiguring VM instance instance-00000008 to attach disk [datastore2] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 544.850022] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2b84fe08-275d-413a-b0ff-5eede530a887 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.867796] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.870273] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 544.870273] env[61768]: value = "task-1228551" [ 544.870273] env[61768]: _type = "Task" [ 544.870273] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.880492] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228551, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.085336] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.152343] env[61768]: INFO nova.compute.manager [-] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Took 1.05 seconds to deallocate network for instance. [ 545.154982] env[61768]: DEBUG nova.compute.claims [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.155281] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.192106] env[61768]: DEBUG oslo_concurrency.lockutils [req-c98f5b11-892d-448c-9f7c-fdccae70159a req-b475712d-b074-49d7-9ed8-8dd546313d13 service nova] Releasing lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.212415] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.361100] env[61768]: DEBUG nova.compute.manager [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Received event network-vif-deleted-47866dda-3b73-4baa-8653-848543575342 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.361327] env[61768]: DEBUG nova.compute.manager [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Received event network-changed-889ff1e9-de2a-434c-9ab7-9a15c77a296b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 545.362164] env[61768]: DEBUG nova.compute.manager [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Refreshing instance network info cache due to event network-changed-889ff1e9-de2a-434c-9ab7-9a15c77a296b. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 545.362164] env[61768]: DEBUG oslo_concurrency.lockutils [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] Acquiring lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.373704] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Releasing lock "refresh_cache-c3b90a4d-e885-453c-9f8e-799adb2400a0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.373954] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 545.374169] env[61768]: DEBUG nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.374315] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.385625] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228551, 'name': ReconfigVM_Task, 'duration_secs': 0.268254} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.385625] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Reconfigured VM instance instance-00000008 to attach disk [datastore2] 0408fe1e-e4fa-4649-83dc-afdd335ee30d/0408fe1e-e4fa-4649-83dc-afdd335ee30d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 545.388562] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e873e6e3-e92c-4ae3-9aa8-1e6eb744fae4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.394853] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 545.394853] env[61768]: value = "task-1228552" [ 545.394853] env[61768]: _type = "Task" [ 545.394853] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.400678] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.408437] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228552, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 545.675681] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cc3bb8-3c40-4b32-915c-c66b1db7cd7d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.683773] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e8c9e41-a42f-4dcf-80d5-6630cf08033a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.714359] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3098143-cdf5-4771-9797-3fb782275408 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.717231] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.717742] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 545.717866] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 545.718092] env[61768]: DEBUG oslo_concurrency.lockutils [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] Acquired lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.718265] env[61768]: DEBUG nova.network.neutron [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Refreshing network info cache for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 545.719265] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-05f5b6ba-2648-4acb-bbe0-1f6143c57052 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.726811] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c874db8-544b-46f1-908f-99b5e2568f48 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.733644] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca7d9256-dd71-4ef2-81ca-1d4143438b37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.754024] env[61768]: DEBUG nova.compute.provider_tree [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.759419] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7bc2075c-2cb2-41c2-9681-780689c1d080 could not be found. [ 545.759691] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 545.759783] env[61768]: INFO nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Took 0.04 seconds to destroy the instance on the hypervisor. [ 545.760038] env[61768]: DEBUG oslo.service.loopingcall [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.760374] env[61768]: DEBUG nova.compute.manager [-] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.760475] env[61768]: DEBUG nova.network.neutron [-] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.783953] env[61768]: DEBUG nova.network.neutron [-] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.906954] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228552, 'name': Rename_Task, 'duration_secs': 0.141567} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.906954] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 545.906954] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c6f87fe-3688-4cb2-8705-89be9fbc60ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.909305] env[61768]: DEBUG nova.network.neutron [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.915019] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Waiting for the task: (returnval){ [ 545.915019] env[61768]: value = "task-1228553" [ 545.915019] env[61768]: _type = "Task" [ 545.915019] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.924174] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228553, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.254139] env[61768]: DEBUG nova.network.neutron [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.261665] env[61768]: DEBUG nova.scheduler.client.report [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.286914] env[61768]: DEBUG nova.network.neutron [-] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.359341] env[61768]: DEBUG nova.network.neutron [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.412399] env[61768]: INFO nova.compute.manager [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: c3b90a4d-e885-453c-9f8e-799adb2400a0] Took 1.04 seconds to deallocate network for instance. [ 546.427311] env[61768]: DEBUG oslo_vmware.api [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Task: {'id': task-1228553, 'name': PowerOnVM_Task, 'duration_secs': 0.439925} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.427671] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 546.428100] env[61768]: DEBUG nova.compute.manager [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 546.429134] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3947fe2-be2f-41c9-8196-c667be74a050 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.767299] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.770019] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.774014] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.698s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.791385] env[61768]: INFO nova.compute.manager [-] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Took 1.03 seconds to deallocate network for instance. [ 546.792230] env[61768]: DEBUG nova.compute.claims [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.792628] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.861999] env[61768]: DEBUG oslo_concurrency.lockutils [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] Releasing lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.861999] env[61768]: DEBUG nova.compute.manager [req-66d50c1d-11c6-43fc-addb-5b0a98151ebc req-ab8c87b8-dc62-40cc-9dd1-5dfa3788584d service nova] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Received event network-vif-deleted-889ff1e9-de2a-434c-9ab7-9a15c77a296b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 546.958891] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.282569] env[61768]: DEBUG nova.compute.utils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.288151] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.288151] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 547.370658] env[61768]: DEBUG nova.policy [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1cf8b3c0383a464982dcc413bc35ef0c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '467ec24b1367459eb83fb0a455970825', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.450516] env[61768]: INFO nova.scheduler.client.report [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Deleted allocations for instance c3b90a4d-e885-453c-9f8e-799adb2400a0 [ 547.570354] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.571146] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.571405] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.571604] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.571769] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.574633] env[61768]: INFO nova.compute.manager [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Terminating instance [ 547.576732] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "refresh_cache-0408fe1e-e4fa-4649-83dc-afdd335ee30d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.577554] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquired lock "refresh_cache-0408fe1e-e4fa-4649-83dc-afdd335ee30d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.577751] env[61768]: DEBUG nova.network.neutron [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.789559] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.852433] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397094ed-7313-49c2-aca6-3f918df3af5e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.864574] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea1e8cf-2dbc-42fc-83b2-01a767d849a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.909036] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca31b6b-aef4-4394-97aa-05d88ddb6ff4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.917144] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0bf6ef-f71a-4810-9fc5-5556096ee2c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.931827] env[61768]: DEBUG nova.compute.provider_tree [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.962615] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d43c0c3e-30d1-4b81-bc9a-dcd48d00202a tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "c3b90a4d-e885-453c-9f8e-799adb2400a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.682s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.085654] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Successfully created port: bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.119393] env[61768]: DEBUG nova.network.neutron [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.234030] env[61768]: DEBUG nova.network.neutron [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.435327] env[61768]: DEBUG nova.scheduler.client.report [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.470254] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 548.737603] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Releasing lock "refresh_cache-0408fe1e-e4fa-4649-83dc-afdd335ee30d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.738244] env[61768]: DEBUG nova.compute.manager [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 548.738584] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 548.739583] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27276f17-8000-4c81-ae63-5f0c8e20a573 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.749317] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 548.749317] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-db675cda-4fe6-419b-9ae0-84c7e82223af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.758498] env[61768]: DEBUG oslo_vmware.api [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 548.758498] env[61768]: value = "task-1228554" [ 548.758498] env[61768]: _type = "Task" [ 548.758498] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.766520] env[61768]: DEBUG oslo_vmware.api [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228554, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.807364] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 548.833777] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:52:30Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='876563391',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-790129883',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.834104] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.834184] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.834876] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.836010] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.836268] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.836548] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.836735] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.836945] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.837174] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.837410] env[61768]: DEBUG nova.virt.hardware [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.838462] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a9422f6-634f-42d5-a210-8aa0aff6070d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.850433] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5f4861-877e-4e59-8c86-b8b50f64b9a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.940269] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.166s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.940926] env[61768]: ERROR nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Traceback (most recent call last): [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self.driver.spawn(context, instance, image_meta, [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] vm_ref = self.build_virtual_machine(instance, [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.940926] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] for vif in network_info: [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return self._sync_wrapper(fn, *args, **kwargs) [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self.wait() [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self[:] = self._gt.wait() [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return self._exit_event.wait() [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] result = hub.switch() [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.941612] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return self.greenlet.switch() [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] result = function(*args, **kwargs) [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] return func(*args, **kwargs) [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] raise e [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] nwinfo = self.network_api.allocate_for_instance( [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] created_port_ids = self._update_ports_for_instance( [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] with excutils.save_and_reraise_exception(): [ 548.942298] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] self.force_reraise() [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] raise self.value [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] updated_port = self._update_port( [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] _ensure_no_port_binding_failure(port) [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] raise exception.PortBindingFailed(port_id=port['id']) [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] nova.exception.PortBindingFailed: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. [ 548.942910] env[61768]: ERROR nova.compute.manager [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] [ 548.946434] env[61768]: DEBUG nova.compute.utils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.946434] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.695s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.946833] env[61768]: INFO nova.compute.claims [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 548.950687] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Build of instance 7d0e752f-9b39-4d01-bedc-5315927deea8 was re-scheduled: Binding failed for port 65e56e3c-761f-4ec3-994c-8b1f396119ed, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.951315] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.951426] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Acquiring lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.951580] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Acquired lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.951752] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 549.002483] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.268702] env[61768]: DEBUG oslo_vmware.api [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228554, 'name': PowerOffVM_Task, 'duration_secs': 0.162954} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.269151] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 549.271051] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 549.271051] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18963cf4-fd85-4601-9a39-0b85358286f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.297237] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 549.297591] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 549.297695] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleting the datastore file [datastore2] 0408fe1e-e4fa-4649-83dc-afdd335ee30d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 549.297972] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-31ad6c02-07df-45d7-afff-7a982ef4af28 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.305207] env[61768]: DEBUG oslo_vmware.api [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for the task: (returnval){ [ 549.305207] env[61768]: value = "task-1228556" [ 549.305207] env[61768]: _type = "Task" [ 549.305207] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.313221] env[61768]: DEBUG oslo_vmware.api [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228556, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.476714] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.576396] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.750151] env[61768]: ERROR nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. [ 549.750151] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.750151] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.750151] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.750151] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.750151] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.750151] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.750151] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.750151] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.750151] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 549.750151] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.750151] env[61768]: ERROR nova.compute.manager raise self.value [ 549.750151] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.750151] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.750151] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.750151] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.750789] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.750789] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.750789] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. [ 549.750789] env[61768]: ERROR nova.compute.manager [ 549.750789] env[61768]: Traceback (most recent call last): [ 549.750789] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.750789] env[61768]: listener.cb(fileno) [ 549.750789] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.750789] env[61768]: result = function(*args, **kwargs) [ 549.750789] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.750789] env[61768]: return func(*args, **kwargs) [ 549.750789] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.750789] env[61768]: raise e [ 549.750789] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.750789] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 549.750789] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.750789] env[61768]: created_port_ids = self._update_ports_for_instance( [ 549.750789] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.750789] env[61768]: with excutils.save_and_reraise_exception(): [ 549.750789] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.750789] env[61768]: self.force_reraise() [ 549.750789] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.750789] env[61768]: raise self.value [ 549.750789] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.750789] env[61768]: updated_port = self._update_port( [ 549.750789] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.750789] env[61768]: _ensure_no_port_binding_failure(port) [ 549.750789] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.750789] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.751789] env[61768]: nova.exception.PortBindingFailed: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. [ 549.751789] env[61768]: Removing descriptor: 17 [ 549.751789] env[61768]: ERROR nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Traceback (most recent call last): [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] yield resources [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self.driver.spawn(context, instance, image_meta, [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.751789] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] vm_ref = self.build_virtual_machine(instance, [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] for vif in network_info: [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return self._sync_wrapper(fn, *args, **kwargs) [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self.wait() [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self[:] = self._gt.wait() [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return self._exit_event.wait() [ 549.752269] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] result = hub.switch() [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return self.greenlet.switch() [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] result = function(*args, **kwargs) [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return func(*args, **kwargs) [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] raise e [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] nwinfo = self.network_api.allocate_for_instance( [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.752714] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] created_port_ids = self._update_ports_for_instance( [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] with excutils.save_and_reraise_exception(): [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self.force_reraise() [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] raise self.value [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] updated_port = self._update_port( [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] _ensure_no_port_binding_failure(port) [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.753171] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] raise exception.PortBindingFailed(port_id=port['id']) [ 549.753660] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] nova.exception.PortBindingFailed: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. [ 549.753660] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] [ 549.755418] env[61768]: INFO nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Terminating instance [ 549.758588] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Acquiring lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.758588] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Acquired lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.758588] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 549.818280] env[61768]: DEBUG oslo_vmware.api [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Task: {'id': task-1228556, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.102179} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.819286] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 549.819286] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 549.819286] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 549.819468] env[61768]: INFO nova.compute.manager [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Took 1.08 seconds to destroy the instance on the hypervisor. [ 549.819727] env[61768]: DEBUG oslo.service.loopingcall [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 549.822053] env[61768]: DEBUG nova.compute.manager [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.822053] env[61768]: DEBUG nova.network.neutron [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 549.847879] env[61768]: DEBUG nova.network.neutron [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.081230] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Releasing lock "refresh_cache-7d0e752f-9b39-4d01-bedc-5315927deea8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.081230] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 550.081390] env[61768]: DEBUG nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.081436] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.104662] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.145631] env[61768]: DEBUG nova.compute.manager [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Received event network-changed-bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 550.145631] env[61768]: DEBUG nova.compute.manager [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Refreshing instance network info cache due to event network-changed-bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 550.145631] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] Acquiring lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.279483] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.317041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Acquiring lock "aa0833ae-99c7-426d-a972-8183667a48f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.317285] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Lock "aa0833ae-99c7-426d-a972-8183667a48f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.352099] env[61768]: DEBUG nova.network.neutron [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.390128] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.435338] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e30b17-2a6c-4fdb-a97d-0402619d2c95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.443104] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cff52e-2d0f-4846-8f78-55fcdda529e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.473982] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c92b94-5537-4f16-908f-69ad6aceb03b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.481078] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bf6fbb-8cde-445f-a616-44017c63fb5b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.494480] env[61768]: DEBUG nova.compute.provider_tree [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.611071] env[61768]: DEBUG nova.network.neutron [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.636924] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Acquiring lock "6540f636-fdfd-48aa-a7b8-0bbc231741fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.637111] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Lock "6540f636-fdfd-48aa-a7b8-0bbc231741fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.854125] env[61768]: INFO nova.compute.manager [-] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Took 1.03 seconds to deallocate network for instance. [ 550.893642] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Releasing lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.894122] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.894322] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 550.894786] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] Acquired lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.894947] env[61768]: DEBUG nova.network.neutron [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Refreshing network info cache for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 550.896217] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88aca5f2-4db1-49ba-b37d-307c4bd79eb5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.910783] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e57222-ca29-4341-b2f8-821e0ac6cde4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.933344] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca31b2ec-0bfb-47d9-9f35-c54659fa7799 could not be found. [ 550.933564] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 550.933746] env[61768]: INFO nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.933983] env[61768]: DEBUG oslo.service.loopingcall [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.934201] env[61768]: DEBUG nova.compute.manager [-] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.934294] env[61768]: DEBUG nova.network.neutron [-] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.961759] env[61768]: DEBUG nova.network.neutron [-] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.997899] env[61768]: DEBUG nova.scheduler.client.report [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.115535] env[61768]: INFO nova.compute.manager [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] [instance: 7d0e752f-9b39-4d01-bedc-5315927deea8] Took 1.03 seconds to deallocate network for instance. [ 551.275322] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "5da49801-148d-4b26-a0a2-816b1d5f7994" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.275322] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "5da49801-148d-4b26-a0a2-816b1d5f7994" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.364729] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.419273] env[61768]: DEBUG nova.network.neutron [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.463538] env[61768]: DEBUG nova.network.neutron [-] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.506702] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.507559] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 551.515788] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.481s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.515788] env[61768]: INFO nova.compute.claims [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.524500] env[61768]: DEBUG nova.network.neutron [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.970579] env[61768]: INFO nova.compute.manager [-] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Took 1.04 seconds to deallocate network for instance. [ 551.972954] env[61768]: DEBUG nova.compute.claims [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.973204] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.016882] env[61768]: DEBUG nova.compute.utils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 552.019979] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 552.020167] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 552.030764] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ed2250a-2475-4618-9680-e824304a84f1 req-84626231-9b90-4f00-a981-75ece8fb3452 service nova] Releasing lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.116539] env[61768]: DEBUG nova.policy [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66184aa9863c4403bdfb8846c344f89e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '382b18a493614de296d4b0c93bb6c1ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 552.159043] env[61768]: INFO nova.scheduler.client.report [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Deleted allocations for instance 7d0e752f-9b39-4d01-bedc-5315927deea8 [ 552.432160] env[61768]: DEBUG nova.compute.manager [req-1b606e15-69e4-4447-bf8a-f0d4aeebd288 req-44bcce8a-a338-410f-bc27-a2c12a3bc82f service nova] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Received event network-vif-deleted-bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.523901] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 552.677565] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b0db6d3b-8dbc-4b04-8e2d-b6a37edb4856 tempest-ServerExternalEventsTest-1760753287 tempest-ServerExternalEventsTest-1760753287-project-member] Lock "7d0e752f-9b39-4d01-bedc-5315927deea8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.885s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.845589] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Successfully created port: 63b50c98-ed09-45dc-bd00-038b07b6f89e {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 553.011077] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b141554a-b0ac-4003-a48e-2984f5938422 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.019287] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea9926d1-19ef-4190-a1ef-c4db2a47bc0b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.054723] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ece9a4-b3ad-4c2e-b333-557284dc1ee5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.065017] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c70a44-040d-45a5-988e-1bad6c2bd07a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.076126] env[61768]: DEBUG nova.compute.provider_tree [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.182715] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.559893] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 553.580802] env[61768]: DEBUG nova.scheduler.client.report [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.594523] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.594523] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.594523] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.594686] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.594686] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.594686] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.594686] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.594686] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.594848] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.594848] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.595520] env[61768]: DEBUG nova.virt.hardware [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.597535] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3441da8e-9b88-4f19-8c13-0f6d591f023f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.607741] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa595c9c-87d8-4c19-aeec-7ac3f7ffdf8c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.721093] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.086945] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.577s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.088225] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.090685] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.634s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.471816] env[61768]: DEBUG nova.compute.manager [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Received event network-changed-63b50c98-ed09-45dc-bd00-038b07b6f89e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 554.471816] env[61768]: DEBUG nova.compute.manager [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Refreshing instance network info cache due to event network-changed-63b50c98-ed09-45dc-bd00-038b07b6f89e. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 554.471956] env[61768]: DEBUG oslo_concurrency.lockutils [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] Acquiring lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.472068] env[61768]: DEBUG oslo_concurrency.lockutils [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] Acquired lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.472237] env[61768]: DEBUG nova.network.neutron [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Refreshing network info cache for port 63b50c98-ed09-45dc-bd00-038b07b6f89e {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 554.550065] env[61768]: ERROR nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. [ 554.550065] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.550065] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.550065] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.550065] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.550065] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.550065] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.550065] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.550065] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.550065] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 554.550065] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.550065] env[61768]: ERROR nova.compute.manager raise self.value [ 554.550065] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.550065] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.550065] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.550065] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.550656] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.550656] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.550656] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. [ 554.550656] env[61768]: ERROR nova.compute.manager [ 554.550656] env[61768]: Traceback (most recent call last): [ 554.550656] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.550656] env[61768]: listener.cb(fileno) [ 554.550656] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.550656] env[61768]: result = function(*args, **kwargs) [ 554.550656] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.550656] env[61768]: return func(*args, **kwargs) [ 554.550656] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.550656] env[61768]: raise e [ 554.550656] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.550656] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 554.550656] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.550656] env[61768]: created_port_ids = self._update_ports_for_instance( [ 554.550656] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.550656] env[61768]: with excutils.save_and_reraise_exception(): [ 554.550656] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.550656] env[61768]: self.force_reraise() [ 554.550656] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.550656] env[61768]: raise self.value [ 554.550656] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.550656] env[61768]: updated_port = self._update_port( [ 554.550656] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.550656] env[61768]: _ensure_no_port_binding_failure(port) [ 554.550656] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.550656] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.551821] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. [ 554.551821] env[61768]: Removing descriptor: 17 [ 554.551821] env[61768]: ERROR nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Traceback (most recent call last): [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] yield resources [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self.driver.spawn(context, instance, image_meta, [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.551821] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] vm_ref = self.build_virtual_machine(instance, [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] for vif in network_info: [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return self._sync_wrapper(fn, *args, **kwargs) [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self.wait() [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self[:] = self._gt.wait() [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return self._exit_event.wait() [ 554.552208] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] result = hub.switch() [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return self.greenlet.switch() [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] result = function(*args, **kwargs) [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return func(*args, **kwargs) [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] raise e [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] nwinfo = self.network_api.allocate_for_instance( [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.552701] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] created_port_ids = self._update_ports_for_instance( [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] with excutils.save_and_reraise_exception(): [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self.force_reraise() [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] raise self.value [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] updated_port = self._update_port( [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] _ensure_no_port_binding_failure(port) [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.553133] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] raise exception.PortBindingFailed(port_id=port['id']) [ 554.553499] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] nova.exception.PortBindingFailed: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. [ 554.553499] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] [ 554.553499] env[61768]: INFO nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Terminating instance [ 554.557365] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Acquiring lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.595905] env[61768]: DEBUG nova.compute.utils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.601134] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 554.601411] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 554.680848] env[61768]: DEBUG nova.policy [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '66030cd94de04a74827a972e2b7c2e2e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '084f2788b08b4b278bc639bf2ad5875c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.958767] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.959147] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.018711] env[61768]: DEBUG nova.network.neutron [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.103684] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.182943] env[61768]: DEBUG nova.network.neutron [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.191331] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6af9b4b-6148-44ca-8358-ccfd1bd52cb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.204798] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1e1067-4cae-4f1b-a930-93e71df0f73a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.238041] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24f1f9e7-0199-49ca-8d3f-afbe716cd2d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.241397] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Successfully created port: f72ba884-e8b4-4f5f-b8b1-d4f475752430 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.248553] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9fb3dcc-b36a-4bff-8c0c-280656f636db {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.263918] env[61768]: DEBUG nova.compute.provider_tree [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.688752] env[61768]: DEBUG oslo_concurrency.lockutils [req-84e61ce2-9765-424f-b04a-eb1ac6749ed2 req-4ec569e0-5079-414c-b564-b4fa7df5cc37 service nova] Releasing lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.690842] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Acquired lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.690842] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 555.770023] env[61768]: DEBUG nova.scheduler.client.report [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.126920] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.153392] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.153659] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.153813] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.153994] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.154208] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.154289] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.154502] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.154755] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.154943] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.155164] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.155327] env[61768]: DEBUG nova.virt.hardware [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.156251] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc344a3d-fec1-4658-8f6d-64fd0f573e35 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.164312] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96dc1d0d-e02e-4f50-a6cd-ba242e024a33 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.229663] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.274625] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.184s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.275413] env[61768]: ERROR nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] Traceback (most recent call last): [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self.driver.spawn(context, instance, image_meta, [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] vm_ref = self.build_virtual_machine(instance, [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.275413] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] for vif in network_info: [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return self._sync_wrapper(fn, *args, **kwargs) [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self.wait() [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self[:] = self._gt.wait() [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return self._exit_event.wait() [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] result = hub.switch() [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.275839] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return self.greenlet.switch() [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] result = function(*args, **kwargs) [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] return func(*args, **kwargs) [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] raise e [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] nwinfo = self.network_api.allocate_for_instance( [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] created_port_ids = self._update_ports_for_instance( [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] with excutils.save_and_reraise_exception(): [ 556.276289] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] self.force_reraise() [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] raise self.value [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] updated_port = self._update_port( [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] _ensure_no_port_binding_failure(port) [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] raise exception.PortBindingFailed(port_id=port['id']) [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] nova.exception.PortBindingFailed: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. [ 556.276697] env[61768]: ERROR nova.compute.manager [instance: 1b694ac3-385b-4083-a35c-542853354202] [ 556.277063] env[61768]: DEBUG nova.compute.utils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 556.277633] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.953s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.279073] env[61768]: INFO nova.compute.claims [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 556.281718] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Build of instance 1b694ac3-385b-4083-a35c-542853354202 was re-scheduled: Binding failed for port 9888a455-daa3-416d-a640-d0ce412d5c77, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 556.282172] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 556.282404] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Acquiring lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.282552] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Acquired lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.282710] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.538133] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.560468] env[61768]: DEBUG nova.compute.manager [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Received event network-vif-deleted-63b50c98-ed09-45dc-bd00-038b07b6f89e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 556.560694] env[61768]: DEBUG nova.compute.manager [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Received event network-changed-f72ba884-e8b4-4f5f-b8b1-d4f475752430 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 556.560841] env[61768]: DEBUG nova.compute.manager [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Refreshing instance network info cache due to event network-changed-f72ba884-e8b4-4f5f-b8b1-d4f475752430. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 556.561260] env[61768]: DEBUG oslo_concurrency.lockutils [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] Acquiring lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.561456] env[61768]: DEBUG oslo_concurrency.lockutils [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] Acquired lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.561629] env[61768]: DEBUG nova.network.neutron [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Refreshing network info cache for port f72ba884-e8b4-4f5f-b8b1-d4f475752430 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 556.765053] env[61768]: ERROR nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. [ 556.765053] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 556.765053] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.765053] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 556.765053] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.765053] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 556.765053] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.765053] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 556.765053] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.765053] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 556.765053] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.765053] env[61768]: ERROR nova.compute.manager raise self.value [ 556.765053] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.765053] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 556.765053] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.765053] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 556.765689] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.765689] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 556.765689] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. [ 556.765689] env[61768]: ERROR nova.compute.manager [ 556.765689] env[61768]: Traceback (most recent call last): [ 556.765689] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 556.765689] env[61768]: listener.cb(fileno) [ 556.765689] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.765689] env[61768]: result = function(*args, **kwargs) [ 556.765689] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.765689] env[61768]: return func(*args, **kwargs) [ 556.765689] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.765689] env[61768]: raise e [ 556.765689] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.765689] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 556.765689] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.765689] env[61768]: created_port_ids = self._update_ports_for_instance( [ 556.765689] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.765689] env[61768]: with excutils.save_and_reraise_exception(): [ 556.765689] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.765689] env[61768]: self.force_reraise() [ 556.765689] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.765689] env[61768]: raise self.value [ 556.765689] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.765689] env[61768]: updated_port = self._update_port( [ 556.765689] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.765689] env[61768]: _ensure_no_port_binding_failure(port) [ 556.765689] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.765689] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 556.766636] env[61768]: nova.exception.PortBindingFailed: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. [ 556.766636] env[61768]: Removing descriptor: 17 [ 556.766636] env[61768]: ERROR nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Traceback (most recent call last): [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] yield resources [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self.driver.spawn(context, instance, image_meta, [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.766636] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] vm_ref = self.build_virtual_machine(instance, [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] for vif in network_info: [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return self._sync_wrapper(fn, *args, **kwargs) [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self.wait() [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self[:] = self._gt.wait() [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return self._exit_event.wait() [ 556.767055] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] result = hub.switch() [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return self.greenlet.switch() [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] result = function(*args, **kwargs) [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return func(*args, **kwargs) [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] raise e [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] nwinfo = self.network_api.allocate_for_instance( [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.767485] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] created_port_ids = self._update_ports_for_instance( [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] with excutils.save_and_reraise_exception(): [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self.force_reraise() [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] raise self.value [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] updated_port = self._update_port( [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] _ensure_no_port_binding_failure(port) [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.767938] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] raise exception.PortBindingFailed(port_id=port['id']) [ 556.768350] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] nova.exception.PortBindingFailed: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. [ 556.768350] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] [ 556.768350] env[61768]: INFO nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Terminating instance [ 556.769046] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Acquiring lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.811658] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.985191] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.043638] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Releasing lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.044173] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 557.044325] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 557.044612] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-49e74c5b-c8ad-4b6e-a6a2-0eccd2d20d37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.055144] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e08915-2951-4f2f-a564-9ef8790de3a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.087988] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7 could not be found. [ 557.087988] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 557.087988] env[61768]: INFO nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 557.087988] env[61768]: DEBUG oslo.service.loopingcall [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.087988] env[61768]: DEBUG nova.compute.manager [-] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.087988] env[61768]: DEBUG nova.network.neutron [-] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.093222] env[61768]: DEBUG nova.network.neutron [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.273161] env[61768]: DEBUG nova.network.neutron [-] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.404523] env[61768]: DEBUG nova.network.neutron [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.492274] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Releasing lock "refresh_cache-1b694ac3-385b-4083-a35c-542853354202" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.494553] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.494553] env[61768]: DEBUG nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.494553] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.516472] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.780412] env[61768]: DEBUG nova.network.neutron [-] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.803160] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402f56bc-557f-47eb-8cce-d873dfa31de6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.815021] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc286cd-548b-45e6-ad73-d1672a276e85 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.848592] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6f6117-0890-4506-aea7-ec7a71db6a25 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.857914] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3acbac2-3c33-4782-a392-46717c63947d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.871998] env[61768]: DEBUG nova.compute.provider_tree [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.910828] env[61768]: DEBUG oslo_concurrency.lockutils [req-5e003f79-4023-49cb-8442-ce0494ef9f2b req-e0849735-ef48-473d-95ad-72a643a265f6 service nova] Releasing lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.910828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Acquired lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.910828] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 558.019498] env[61768]: DEBUG nova.network.neutron [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.285677] env[61768]: INFO nova.compute.manager [-] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Took 1.20 seconds to deallocate network for instance. [ 558.288257] env[61768]: DEBUG nova.compute.claims [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 558.289324] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.378895] env[61768]: DEBUG nova.scheduler.client.report [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 558.439863] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.528148] env[61768]: INFO nova.compute.manager [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] [instance: 1b694ac3-385b-4083-a35c-542853354202] Took 1.03 seconds to deallocate network for instance. [ 558.566029] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.643552] env[61768]: DEBUG nova.compute.manager [req-8415e3cc-243e-49fd-860c-0632850b358f req-966eae60-aa53-4219-99ee-c20d114779c5 service nova] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Received event network-vif-deleted-f72ba884-e8b4-4f5f-b8b1-d4f475752430 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 558.884223] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.885273] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 558.890220] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.942s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.070993] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Releasing lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.071720] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.071720] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 559.072106] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-83bcf822-5ad3-4849-8e42-bdc17ba6f70a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.088019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d768c0f-84f8-4be2-adc2-72070a2a0fdb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.112362] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1331a447-7be8-42b8-9d59-6b97742e49d1 could not be found. [ 559.112760] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 559.113125] env[61768]: INFO nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.113549] env[61768]: DEBUG oslo.service.loopingcall [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.113912] env[61768]: DEBUG nova.compute.manager [-] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.114146] env[61768]: DEBUG nova.network.neutron [-] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.131408] env[61768]: DEBUG nova.network.neutron [-] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.399469] env[61768]: DEBUG nova.compute.utils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 559.401894] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 559.402094] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 559.478764] env[61768]: DEBUG nova.policy [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a50259d0cd5468298a107ee352d4c21', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5e793b264e244c988a96ab7b27b7c967', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 559.563361] env[61768]: INFO nova.scheduler.client.report [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Deleted allocations for instance 1b694ac3-385b-4083-a35c-542853354202 [ 559.640326] env[61768]: DEBUG nova.network.neutron [-] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.832187] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Successfully created port: 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 559.906712] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 559.967751] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7954446-8ad6-4606-8cd0-a4a41e4736d7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.975982] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8035a40d-86e8-493d-8aa3-117adcb1c2d1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.011497] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443ab0ea-ea25-45bb-b00d-45a76dc59a28 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.019647] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1dbe29-7b5b-4c1a-8bcf-c19f19100068 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.035015] env[61768]: DEBUG nova.compute.provider_tree [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 560.080439] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ccfada1-4065-4a67-8a5a-282df9135d65 tempest-VolumesAssistedSnapshotsTest-360971385 tempest-VolumesAssistedSnapshotsTest-360971385-project-member] Lock "1b694ac3-385b-4083-a35c-542853354202" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.063s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.142493] env[61768]: INFO nova.compute.manager [-] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Took 1.03 seconds to deallocate network for instance. [ 560.145297] env[61768]: DEBUG nova.compute.claims [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.145489] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.540216] env[61768]: DEBUG nova.scheduler.client.report [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.587856] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.920762] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 560.956704] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 560.957035] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 560.957255] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.957539] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 560.957749] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.957947] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 560.958970] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 560.959575] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 560.959831] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 560.960046] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 560.960778] env[61768]: DEBUG nova.virt.hardware [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 560.961318] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87a164d-72d1-46dc-81d9-7815b6b73b89 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.970560] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b483c7f3-a951-4597-85e5-2ee5d79424f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.048520] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.156s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 561.048520] env[61768]: ERROR nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. [ 561.048520] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Traceback (most recent call last): [ 561.048520] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.048520] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self.driver.spawn(context, instance, image_meta, [ 561.048520] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.048520] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.048520] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.048520] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] vm_ref = self.build_virtual_machine(instance, [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] for vif in network_info: [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return self._sync_wrapper(fn, *args, **kwargs) [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self.wait() [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self[:] = self._gt.wait() [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return self._exit_event.wait() [ 561.048897] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] result = hub.switch() [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return self.greenlet.switch() [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] result = function(*args, **kwargs) [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] return func(*args, **kwargs) [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] raise e [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] nwinfo = self.network_api.allocate_for_instance( [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.049288] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] created_port_ids = self._update_ports_for_instance( [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] with excutils.save_and_reraise_exception(): [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] self.force_reraise() [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] raise self.value [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] updated_port = self._update_port( [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] _ensure_no_port_binding_failure(port) [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.049703] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] raise exception.PortBindingFailed(port_id=port['id']) [ 561.050046] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] nova.exception.PortBindingFailed: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. [ 561.050046] env[61768]: ERROR nova.compute.manager [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] [ 561.050046] env[61768]: DEBUG nova.compute.utils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 561.050046] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 25.932s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.050046] env[61768]: DEBUG nova.objects.instance [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lazy-loading 'resources' on Instance uuid ca465e62-c190-4d67-9dec-87ea1aa8f194 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 561.051482] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Build of instance 417480ad-61d9-447d-a84d-0cf5df94f4bc was re-scheduled: Binding failed for port 0591a789-7773-484e-9663-e763b0bb4667, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 561.051923] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 561.052170] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.052325] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.052485] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.122129] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.223498] env[61768]: ERROR nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. [ 561.223498] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 561.223498] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.223498] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 561.223498] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.223498] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 561.223498] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.223498] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 561.223498] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.223498] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 561.223498] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.223498] env[61768]: ERROR nova.compute.manager raise self.value [ 561.223498] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.223498] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 561.223498] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.223498] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 561.224932] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.224932] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 561.224932] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. [ 561.224932] env[61768]: ERROR nova.compute.manager [ 561.224932] env[61768]: Traceback (most recent call last): [ 561.224932] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 561.224932] env[61768]: listener.cb(fileno) [ 561.224932] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.224932] env[61768]: result = function(*args, **kwargs) [ 561.224932] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.224932] env[61768]: return func(*args, **kwargs) [ 561.224932] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.224932] env[61768]: raise e [ 561.224932] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.224932] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 561.224932] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.224932] env[61768]: created_port_ids = self._update_ports_for_instance( [ 561.224932] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.224932] env[61768]: with excutils.save_and_reraise_exception(): [ 561.224932] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.224932] env[61768]: self.force_reraise() [ 561.224932] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.224932] env[61768]: raise self.value [ 561.224932] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.224932] env[61768]: updated_port = self._update_port( [ 561.224932] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.224932] env[61768]: _ensure_no_port_binding_failure(port) [ 561.224932] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.224932] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 561.226180] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. [ 561.226180] env[61768]: Removing descriptor: 17 [ 561.226180] env[61768]: ERROR nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Traceback (most recent call last): [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] yield resources [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self.driver.spawn(context, instance, image_meta, [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 561.226180] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] vm_ref = self.build_virtual_machine(instance, [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] for vif in network_info: [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return self._sync_wrapper(fn, *args, **kwargs) [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self.wait() [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self[:] = self._gt.wait() [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return self._exit_event.wait() [ 561.226525] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] result = hub.switch() [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return self.greenlet.switch() [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] result = function(*args, **kwargs) [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return func(*args, **kwargs) [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] raise e [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] nwinfo = self.network_api.allocate_for_instance( [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 561.226933] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] created_port_ids = self._update_ports_for_instance( [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] with excutils.save_and_reraise_exception(): [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self.force_reraise() [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] raise self.value [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] updated_port = self._update_port( [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] _ensure_no_port_binding_failure(port) [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 561.227320] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] raise exception.PortBindingFailed(port_id=port['id']) [ 561.228166] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] nova.exception.PortBindingFailed: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. [ 561.228166] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] [ 561.228166] env[61768]: INFO nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Terminating instance [ 561.228166] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Acquiring lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.228166] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Acquired lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.228166] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 561.277730] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Acquiring lock "c299abc5-7913-4eea-ac0a-02e713c6009c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.277962] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Lock "c299abc5-7913-4eea-ac0a-02e713c6009c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.278869] env[61768]: DEBUG nova.compute.manager [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Received event network-changed-2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 561.279255] env[61768]: DEBUG nova.compute.manager [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Refreshing instance network info cache due to event network-changed-2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 561.279434] env[61768]: DEBUG oslo_concurrency.lockutils [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] Acquiring lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.577713] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.664918] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.753521] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.001999] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-538b4baf-7b58-49d1-8e46-c46d245c40f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.013032] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3715637e-4004-413d-88d4-40de48a0b2a5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.045635] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10ac82c-8ce7-4f44-9014-d46af968151d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.053185] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c78fead-6188-4b28-b122-b4dd106332c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.065809] env[61768]: DEBUG nova.compute.provider_tree [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.067607] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.168431] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-417480ad-61d9-447d-a84d-0cf5df94f4bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.168754] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 562.169016] env[61768]: DEBUG nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.169551] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 562.185658] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.571074] env[61768]: DEBUG nova.scheduler.client.report [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.575778] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Releasing lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.576176] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 562.576365] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 562.576892] env[61768]: DEBUG oslo_concurrency.lockutils [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] Acquired lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.577076] env[61768]: DEBUG nova.network.neutron [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Refreshing network info cache for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 562.578046] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41faef63-5254-4c2d-b6b2-0a1c21f00211 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.594213] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8e17f39-123e-459b-ab0a-009e6b47df9d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.616747] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1b9b618-91b5-429b-8581-9f819e0d81ee could not be found. [ 562.619106] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 562.619106] env[61768]: INFO nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Took 0.04 seconds to destroy the instance on the hypervisor. [ 562.619106] env[61768]: DEBUG oslo.service.loopingcall [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 562.619106] env[61768]: DEBUG nova.compute.manager [-] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 562.619106] env[61768]: DEBUG nova.network.neutron [-] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 562.643555] env[61768]: DEBUG nova.network.neutron [-] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.689723] env[61768]: DEBUG nova.network.neutron [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.081274] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.028s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.081274] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 26.272s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.081274] env[61768]: DEBUG nova.objects.instance [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 563.111117] env[61768]: DEBUG nova.network.neutron [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 563.115296] env[61768]: INFO nova.scheduler.client.report [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Deleted allocations for instance ca465e62-c190-4d67-9dec-87ea1aa8f194 [ 563.144879] env[61768]: DEBUG nova.network.neutron [-] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.194035] env[61768]: INFO nova.compute.manager [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 417480ad-61d9-447d-a84d-0cf5df94f4bc] Took 1.02 seconds to deallocate network for instance. [ 563.300984] env[61768]: DEBUG nova.compute.manager [req-cdf577bf-162e-431e-af18-280f306ec2c1 req-49fe2f5c-129e-4129-aa6f-d6df6d4f044b service nova] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Received event network-vif-deleted-2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.310708] env[61768]: DEBUG nova.network.neutron [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.629519] env[61768]: DEBUG oslo_concurrency.lockutils [None req-26c0b542-75b1-42af-b953-e8584e8bc96e tempest-ServerDiagnosticsV248Test-1968214408 tempest-ServerDiagnosticsV248Test-1968214408-project-member] Lock "ca465e62-c190-4d67-9dec-87ea1aa8f194" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.494s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.648516] env[61768]: INFO nova.compute.manager [-] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Took 1.03 seconds to deallocate network for instance. [ 563.656611] env[61768]: DEBUG nova.compute.claims [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 563.656806] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.815299] env[61768]: DEBUG oslo_concurrency.lockutils [req-7deb3840-5256-4613-ae73-78f127a96b34 req-7a577009-5913-40a2-9356-63ceca635a51 service nova] Releasing lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 564.101840] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cc0c12a6-8186-48e1-84ab-02250e71866d tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.104020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.948s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.243494] env[61768]: INFO nova.scheduler.client.report [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleted allocations for instance 417480ad-61d9-447d-a84d-0cf5df94f4bc [ 564.756198] env[61768]: DEBUG oslo_concurrency.lockutils [None req-119c344b-16fc-4070-8253-5f4f27050b31 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "417480ad-61d9-447d-a84d-0cf5df94f4bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.836s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.138038] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf91e6b2-0463-4438-aa67-321fcfbea8ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.145786] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c90210-a22e-48ec-9b91-250e08586106 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.177151] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdf687e-56b9-4905-b3f4-873baddc0eb8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.189351] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ace84e51-7ac5-47a4-9a79-b8784ba2d64f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.202788] env[61768]: DEBUG nova.compute.provider_tree [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.260624] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 565.706045] env[61768]: DEBUG nova.scheduler.client.report [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.789104] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.212472] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.108s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.214279] env[61768]: ERROR nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Traceback (most recent call last): [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self.driver.spawn(context, instance, image_meta, [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] vm_ref = self.build_virtual_machine(instance, [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] vif_infos = vmwarevif.get_vif_info(self._session, [ 566.214279] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] for vif in network_info: [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return self._sync_wrapper(fn, *args, **kwargs) [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self.wait() [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self[:] = self._gt.wait() [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return self._exit_event.wait() [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] result = hub.switch() [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 566.214696] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return self.greenlet.switch() [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] result = function(*args, **kwargs) [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] return func(*args, **kwargs) [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] raise e [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] nwinfo = self.network_api.allocate_for_instance( [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] created_port_ids = self._update_ports_for_instance( [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] with excutils.save_and_reraise_exception(): [ 566.215212] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] self.force_reraise() [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] raise self.value [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] updated_port = self._update_port( [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] _ensure_no_port_binding_failure(port) [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] raise exception.PortBindingFailed(port_id=port['id']) [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] nova.exception.PortBindingFailed: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. [ 566.215688] env[61768]: ERROR nova.compute.manager [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] [ 566.216045] env[61768]: DEBUG nova.compute.utils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 566.217537] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.425s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.222507] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Build of instance eb8203da-3427-48a8-9ec0-a9d12d8c44d2 was re-scheduled: Binding failed for port 47866dda-3b73-4baa-8653-848543575342, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 566.223075] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 566.223318] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.223486] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquired lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.223652] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 566.763308] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.825114] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Acquiring lock "a83f5b0f-508c-4829-a4a6-641baa7b4c95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.825407] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Lock "a83f5b0f-508c-4829-a4a6-641baa7b4c95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.869570] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.225027] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de33285-c0cc-45bb-86a1-57ab0775128d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.232544] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4192ee2-45c5-4d75-9755-84c2216eedc9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.264432] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f2c852-bcc6-4e76-a729-b9cefc7de752 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.272105] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c52fd9-0120-4b0a-8c5d-9ddd6e557bdc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.290922] env[61768]: DEBUG nova.compute.provider_tree [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.372500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Releasing lock "refresh_cache-eb8203da-3427-48a8-9ec0-a9d12d8c44d2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.375522] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 567.375522] env[61768]: DEBUG nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 567.375522] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.389486] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.794080] env[61768]: DEBUG nova.scheduler.client.report [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.805220] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Acquiring lock "f6b07607-632d-46d9-a72f-5a524fd20a3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.805220] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Lock "f6b07607-632d-46d9-a72f-5a524fd20a3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.892681] env[61768]: DEBUG nova.network.neutron [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.305697] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.088s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.306431] env[61768]: ERROR nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Traceback (most recent call last): [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self.driver.spawn(context, instance, image_meta, [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] vm_ref = self.build_virtual_machine(instance, [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.306431] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] for vif in network_info: [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return self._sync_wrapper(fn, *args, **kwargs) [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self.wait() [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self[:] = self._gt.wait() [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return self._exit_event.wait() [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] result = hub.switch() [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.306784] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return self.greenlet.switch() [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] result = function(*args, **kwargs) [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] return func(*args, **kwargs) [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] raise e [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] nwinfo = self.network_api.allocate_for_instance( [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] created_port_ids = self._update_ports_for_instance( [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] with excutils.save_and_reraise_exception(): [ 568.307168] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] self.force_reraise() [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] raise self.value [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] updated_port = self._update_port( [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] _ensure_no_port_binding_failure(port) [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] raise exception.PortBindingFailed(port_id=port['id']) [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] nova.exception.PortBindingFailed: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. [ 568.307614] env[61768]: ERROR nova.compute.manager [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] [ 568.307957] env[61768]: DEBUG nova.compute.utils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 568.308352] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 21.350s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.308528] env[61768]: DEBUG nova.objects.instance [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 568.310970] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Build of instance 7bc2075c-2cb2-41c2-9681-780689c1d080 was re-scheduled: Binding failed for port 889ff1e9-de2a-434c-9ab7-9a15c77a296b, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 568.311426] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 568.311647] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.311791] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.311946] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 568.395750] env[61768]: INFO nova.compute.manager [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: eb8203da-3427-48a8-9ec0-a9d12d8c44d2] Took 1.02 seconds to deallocate network for instance. [ 568.763035] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.763292] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "12a543c2-1081-49d7-800b-07f0a2516904" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.843996] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.999014] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.320395] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aca8b87e-aa9d-426c-a566-35edbba0139c tempest-ServersAdmin275Test-12035787 tempest-ServersAdmin275Test-12035787-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.322209] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.320s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.323196] env[61768]: INFO nova.compute.claims [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 569.437902] env[61768]: INFO nova.scheduler.client.report [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Deleted allocations for instance eb8203da-3427-48a8-9ec0-a9d12d8c44d2 [ 569.502572] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "refresh_cache-7bc2075c-2cb2-41c2-9681-780689c1d080" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.502572] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 569.502774] env[61768]: DEBUG nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 569.502774] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 569.523242] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.944687] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5923db3b-7431-4f80-9c3a-6ef8833497d4 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "eb8203da-3427-48a8-9ec0-a9d12d8c44d2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.469s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.027676] env[61768]: DEBUG nova.network.neutron [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.449698] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.532906] env[61768]: INFO nova.compute.manager [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 7bc2075c-2cb2-41c2-9681-780689c1d080] Took 1.03 seconds to deallocate network for instance. [ 570.778926] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bade49c8-08f9-48bb-b9dd-62f2a2915bb9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.787180] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade017c0-0b41-4ef3-b47c-1304f4db5763 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.820088] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36fbe2a-36d7-468d-8122-a3a5e68c80e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.828063] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df5ac7c-b497-458c-b60d-5d41e16e3aa3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.841350] env[61768]: DEBUG nova.compute.provider_tree [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 570.972970] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.349846] env[61768]: DEBUG nova.scheduler.client.report [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.567029] env[61768]: INFO nova.scheduler.client.report [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted allocations for instance 7bc2075c-2cb2-41c2-9681-780689c1d080 [ 571.854167] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.530s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.854167] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 571.857318] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.491s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.857318] env[61768]: DEBUG nova.objects.instance [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lazy-loading 'resources' on Instance uuid 0408fe1e-e4fa-4649-83dc-afdd335ee30d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 572.074417] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2b1eaaa2-2897-44c2-859f-4a857a63f6be tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "7bc2075c-2cb2-41c2-9681-780689c1d080" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.204s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.362982] env[61768]: DEBUG nova.compute.utils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.364627] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.364788] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 572.452797] env[61768]: DEBUG nova.policy [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9d7f72b4c22452d89004ca70d165e67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f4804c086b449608ef34ea22c018907', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 572.578033] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.840999] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e6f6e9-47c0-41ab-864d-589fb0f70ab9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.852587] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f524c535-bfdd-4b57-af9a-7fc6a45e85b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.884591] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 572.888157] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f584eef5-91ac-4d11-97c7-25d4dca4e0b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.897450] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7fb2e17-a6da-48b0-9ad6-b2ee8c198e85 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.911944] env[61768]: DEBUG nova.compute.provider_tree [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.105112] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.206262] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Successfully created port: dfbb10fa-71a5-4f65-beff-de178b6f1680 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 573.418326] env[61768]: DEBUG nova.scheduler.client.report [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 573.674459] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.674698] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.899027] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 573.924150] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.068s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.927284] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.953s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.941109] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 573.941377] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 573.941582] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 573.941821] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 573.941965] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 573.942126] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 573.942369] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 573.942569] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 573.942785] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 573.942960] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 573.944020] env[61768]: DEBUG nova.virt.hardware [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 573.944469] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36adf284-ffd5-4f1c-9952-8867e216f3ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.954268] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c3d547-b230-471b-90df-a6b28037a47e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.959333] env[61768]: INFO nova.scheduler.client.report [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Deleted allocations for instance 0408fe1e-e4fa-4649-83dc-afdd335ee30d [ 574.475157] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b338358a-7518-403f-971e-1ff2b2a3cb58 tempest-ServersAdmin275Test-1812171771 tempest-ServersAdmin275Test-1812171771-project-member] Lock "0408fe1e-e4fa-4649-83dc-afdd335ee30d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.904s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.928321] env[61768]: ERROR nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. [ 574.928321] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 574.928321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.928321] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 574.928321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.928321] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 574.928321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.928321] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 574.928321] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.928321] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 574.928321] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.928321] env[61768]: ERROR nova.compute.manager raise self.value [ 574.928321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.928321] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 574.928321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.928321] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 574.929122] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.929122] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 574.929122] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. [ 574.929122] env[61768]: ERROR nova.compute.manager [ 574.929122] env[61768]: Traceback (most recent call last): [ 574.929122] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 574.929122] env[61768]: listener.cb(fileno) [ 574.929122] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.929122] env[61768]: result = function(*args, **kwargs) [ 574.929122] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.929122] env[61768]: return func(*args, **kwargs) [ 574.929122] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.929122] env[61768]: raise e [ 574.929122] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.929122] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 574.929122] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.929122] env[61768]: created_port_ids = self._update_ports_for_instance( [ 574.929122] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.929122] env[61768]: with excutils.save_and_reraise_exception(): [ 574.929122] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.929122] env[61768]: self.force_reraise() [ 574.929122] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.929122] env[61768]: raise self.value [ 574.929122] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.929122] env[61768]: updated_port = self._update_port( [ 574.929122] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.929122] env[61768]: _ensure_no_port_binding_failure(port) [ 574.929122] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.929122] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 574.930032] env[61768]: nova.exception.PortBindingFailed: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. [ 574.930032] env[61768]: Removing descriptor: 17 [ 574.930032] env[61768]: ERROR nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Traceback (most recent call last): [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] yield resources [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self.driver.spawn(context, instance, image_meta, [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 574.930032] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] vm_ref = self.build_virtual_machine(instance, [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] vif_infos = vmwarevif.get_vif_info(self._session, [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] for vif in network_info: [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return self._sync_wrapper(fn, *args, **kwargs) [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self.wait() [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self[:] = self._gt.wait() [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return self._exit_event.wait() [ 574.930775] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] result = hub.switch() [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return self.greenlet.switch() [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] result = function(*args, **kwargs) [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return func(*args, **kwargs) [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] raise e [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] nwinfo = self.network_api.allocate_for_instance( [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 574.931246] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] created_port_ids = self._update_ports_for_instance( [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] with excutils.save_and_reraise_exception(): [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self.force_reraise() [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] raise self.value [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] updated_port = self._update_port( [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] _ensure_no_port_binding_failure(port) [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 574.931659] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] raise exception.PortBindingFailed(port_id=port['id']) [ 574.932379] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] nova.exception.PortBindingFailed: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. [ 574.932379] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] [ 574.932379] env[61768]: INFO nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Terminating instance [ 574.933715] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 574.933715] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 574.933715] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 574.958052] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70eaf716-f4f1-439e-bfc5-ab350a83be07 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.968231] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac63e4fe-f9ea-4950-bfe4-23997a0e2780 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.973516] env[61768]: DEBUG nova.compute.manager [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Received event network-changed-dfbb10fa-71a5-4f65-beff-de178b6f1680 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 574.973731] env[61768]: DEBUG nova.compute.manager [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Refreshing instance network info cache due to event network-changed-dfbb10fa-71a5-4f65-beff-de178b6f1680. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 574.973919] env[61768]: DEBUG oslo_concurrency.lockutils [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] Acquiring lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.004663] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dde08476-89d6-4c92-bc11-0f8e10300baa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.013255] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481a248e-c4ce-416d-b90d-b659965e3d31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.026904] env[61768]: DEBUG nova.compute.provider_tree [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 575.453848] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 575.532015] env[61768]: DEBUG nova.scheduler.client.report [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 575.586144] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.035953] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.109s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.036633] env[61768]: ERROR nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Traceback (most recent call last): [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self.driver.spawn(context, instance, image_meta, [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] vm_ref = self.build_virtual_machine(instance, [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.036633] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] for vif in network_info: [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return self._sync_wrapper(fn, *args, **kwargs) [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self.wait() [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self[:] = self._gt.wait() [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return self._exit_event.wait() [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] result = hub.switch() [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.036939] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return self.greenlet.switch() [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] result = function(*args, **kwargs) [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] return func(*args, **kwargs) [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] raise e [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] nwinfo = self.network_api.allocate_for_instance( [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] created_port_ids = self._update_ports_for_instance( [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] with excutils.save_and_reraise_exception(): [ 576.037256] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] self.force_reraise() [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] raise self.value [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] updated_port = self._update_port( [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] _ensure_no_port_binding_failure(port) [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] raise exception.PortBindingFailed(port_id=port['id']) [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] nova.exception.PortBindingFailed: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. [ 576.037584] env[61768]: ERROR nova.compute.manager [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] [ 576.037841] env[61768]: DEBUG nova.compute.utils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 576.038533] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.318s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.040045] env[61768]: INFO nova.compute.claims [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 576.042670] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Build of instance ca31b2ec-0bfb-47d9-9f35-c54659fa7799 was re-scheduled: Binding failed for port bb8e73a0-4cb0-43f2-b8a7-ae3cd81400f3, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 576.043117] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 576.043389] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Acquiring lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.043514] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Acquired lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.043670] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 576.089369] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.089787] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 576.089981] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 576.090330] env[61768]: DEBUG oslo_concurrency.lockutils [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] Acquired lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.090511] env[61768]: DEBUG nova.network.neutron [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Refreshing network info cache for port dfbb10fa-71a5-4f65-beff-de178b6f1680 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 576.092069] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2ffdc6b-7cdf-4d7f-b68a-c6d5cb7bafe7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.102012] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed1bb7ca-bed1-49ac-9345-6a090de1c863 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.125034] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe could not be found. [ 576.125034] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 576.125034] env[61768]: INFO nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Took 0.03 seconds to destroy the instance on the hypervisor. [ 576.125253] env[61768]: DEBUG oslo.service.loopingcall [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.125413] env[61768]: DEBUG nova.compute.manager [-] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.125502] env[61768]: DEBUG nova.network.neutron [-] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.149481] env[61768]: DEBUG nova.network.neutron [-] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.580742] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.641067] env[61768]: DEBUG nova.network.neutron [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.652573] env[61768]: DEBUG nova.network.neutron [-] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.704588] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.746502] env[61768]: DEBUG nova.network.neutron [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.916501] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.916501] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.154992] env[61768]: INFO nova.compute.manager [-] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Took 1.03 seconds to deallocate network for instance. [ 577.158159] env[61768]: DEBUG nova.compute.manager [req-f6876e80-816a-40a3-b798-9b33e8f219ac req-328b05c8-76fb-4883-aa3a-1d4345de761c service nova] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Received event network-vif-deleted-dfbb10fa-71a5-4f65-beff-de178b6f1680 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 577.164209] env[61768]: DEBUG nova.compute.claims [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 577.164448] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.209962] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Releasing lock "refresh_cache-ca31b2ec-0bfb-47d9-9f35-c54659fa7799" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.210273] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 577.210487] env[61768]: DEBUG nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.210681] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 577.237956] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.249093] env[61768]: DEBUG oslo_concurrency.lockutils [req-374c0b8d-7c44-47aa-a57a-a24bead49a08 req-3b35177e-19fc-4bb5-8b3d-1228cb69ef9a service nova] Releasing lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.534672] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff72136-8026-41aa-b868-ee07474913f5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.543889] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d85a8d-c87b-4576-9725-d1c080a307b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.581420] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63117d39-46c8-441a-a17c-685bd98ca6ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.590150] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8783ee22-b0de-453e-a1b1-f16e2b40fa34 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.606970] env[61768]: DEBUG nova.compute.provider_tree [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 577.741162] env[61768]: DEBUG nova.network.neutron [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.113896] env[61768]: DEBUG nova.scheduler.client.report [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 578.246654] env[61768]: INFO nova.compute.manager [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] [instance: ca31b2ec-0bfb-47d9-9f35-c54659fa7799] Took 1.04 seconds to deallocate network for instance. [ 578.618927] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.580s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 578.619904] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 578.622069] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.333s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.126992] env[61768]: DEBUG nova.compute.utils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 579.128443] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 579.128994] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 579.177156] env[61768]: DEBUG nova.policy [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe89eba3bb3141ba9935081bbd49e77a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ccb1589efa094bd8a121fd5817828417', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 579.280099] env[61768]: INFO nova.scheduler.client.report [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Deleted allocations for instance ca31b2ec-0bfb-47d9-9f35-c54659fa7799 [ 579.403823] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "0df161e9-4a66-432e-9366-60f231c53e3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.404090] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "0df161e9-4a66-432e-9366-60f231c53e3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.601395] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Successfully created port: 56711858-18ee-4dbf-9461-8f12bb7ab446 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.624439] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10398ba9-4e6b-4ca0-8a3e-2f8c6615d516 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.634927] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47859a3e-9362-4c6d-b700-fa9bb715f92d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.639163] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 579.678907] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46597c2d-3a3d-4691-8f1b-b575eb8a4e7b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.687196] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e1f15ff-5344-4e2e-a833-557c88de4d8d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.701027] env[61768]: DEBUG nova.compute.provider_tree [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.795255] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5133b8e4-c797-4ac4-b71b-19849697c354 tempest-ServersWithSpecificFlavorTestJSON-812732162 tempest-ServersWithSpecificFlavorTestJSON-812732162-project-member] Lock "ca31b2ec-0bfb-47d9-9f35-c54659fa7799" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.794s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.204676] env[61768]: DEBUG nova.scheduler.client.report [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 580.300631] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.649819] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 580.687680] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 580.687680] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 580.687872] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 580.688182] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 580.688447] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 580.688715] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 580.690168] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 580.690532] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 580.690917] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 580.691290] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 580.691655] env[61768]: DEBUG nova.virt.hardware [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 580.693275] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363b2129-4830-44c5-8001-c02a98592e3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.703066] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-806d66f5-35b1-4e6c-8a2d-a5277f0e7294 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.709724] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.088s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.710437] env[61768]: ERROR nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Traceback (most recent call last): [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self.driver.spawn(context, instance, image_meta, [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] vm_ref = self.build_virtual_machine(instance, [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.710437] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] for vif in network_info: [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return self._sync_wrapper(fn, *args, **kwargs) [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self.wait() [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self[:] = self._gt.wait() [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return self._exit_event.wait() [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] result = hub.switch() [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.710763] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return self.greenlet.switch() [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] result = function(*args, **kwargs) [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] return func(*args, **kwargs) [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] raise e [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] nwinfo = self.network_api.allocate_for_instance( [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] created_port_ids = self._update_ports_for_instance( [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] with excutils.save_and_reraise_exception(): [ 580.711097] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] self.force_reraise() [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] raise self.value [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] updated_port = self._update_port( [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] _ensure_no_port_binding_failure(port) [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] raise exception.PortBindingFailed(port_id=port['id']) [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] nova.exception.PortBindingFailed: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. [ 580.711420] env[61768]: ERROR nova.compute.manager [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] [ 580.711671] env[61768]: DEBUG nova.compute.utils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 580.720781] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.575s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.728870] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Build of instance 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7 was re-scheduled: Binding failed for port 63b50c98-ed09-45dc-bd00-038b07b6f89e, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 580.731226] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 580.731226] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Acquiring lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.731396] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Acquired lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.731612] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.830321] env[61768]: ERROR nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. [ 580.830321] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 580.830321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.830321] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 580.830321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.830321] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 580.830321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.830321] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 580.830321] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.830321] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 580.830321] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.830321] env[61768]: ERROR nova.compute.manager raise self.value [ 580.830321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.830321] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 580.830321] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.830321] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 580.830825] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.830825] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 580.830825] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. [ 580.830825] env[61768]: ERROR nova.compute.manager [ 580.830825] env[61768]: Traceback (most recent call last): [ 580.830825] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 580.830825] env[61768]: listener.cb(fileno) [ 580.830825] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.830825] env[61768]: result = function(*args, **kwargs) [ 580.830825] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.830825] env[61768]: return func(*args, **kwargs) [ 580.830825] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.830825] env[61768]: raise e [ 580.830825] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.830825] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 580.830825] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.830825] env[61768]: created_port_ids = self._update_ports_for_instance( [ 580.830825] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.830825] env[61768]: with excutils.save_and_reraise_exception(): [ 580.830825] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.830825] env[61768]: self.force_reraise() [ 580.830825] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.830825] env[61768]: raise self.value [ 580.830825] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.830825] env[61768]: updated_port = self._update_port( [ 580.830825] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.830825] env[61768]: _ensure_no_port_binding_failure(port) [ 580.830825] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.830825] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 580.831620] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. [ 580.831620] env[61768]: Removing descriptor: 14 [ 580.831620] env[61768]: ERROR nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Traceback (most recent call last): [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] yield resources [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self.driver.spawn(context, instance, image_meta, [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.831620] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] vm_ref = self.build_virtual_machine(instance, [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] for vif in network_info: [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return self._sync_wrapper(fn, *args, **kwargs) [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self.wait() [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self[:] = self._gt.wait() [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return self._exit_event.wait() [ 580.831920] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] result = hub.switch() [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return self.greenlet.switch() [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] result = function(*args, **kwargs) [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return func(*args, **kwargs) [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] raise e [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] nwinfo = self.network_api.allocate_for_instance( [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.832248] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] created_port_ids = self._update_ports_for_instance( [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] with excutils.save_and_reraise_exception(): [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self.force_reraise() [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] raise self.value [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] updated_port = self._update_port( [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] _ensure_no_port_binding_failure(port) [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.832562] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] raise exception.PortBindingFailed(port_id=port['id']) [ 580.832858] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] nova.exception.PortBindingFailed: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. [ 580.832858] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] [ 580.832858] env[61768]: INFO nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Terminating instance [ 580.833945] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Acquiring lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.834075] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Acquired lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.834881] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.837729] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.887594] env[61768]: DEBUG nova.compute.manager [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Received event network-changed-56711858-18ee-4dbf-9461-8f12bb7ab446 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.887982] env[61768]: DEBUG nova.compute.manager [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Refreshing instance network info cache due to event network-changed-56711858-18ee-4dbf-9461-8f12bb7ab446. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 580.888289] env[61768]: DEBUG oslo_concurrency.lockutils [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] Acquiring lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.262042] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.361817] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.397057] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.486975] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.644007] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a46973-8bf5-4ead-a443-777584e92291 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.651651] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3526d503-2273-4c65-b136-6b67150c97ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.682458] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfb47201-04f8-4bed-aab9-9c6f5c6f0344 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.690583] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47bc53d7-ec95-4cf9-ac12-374a874ba245 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.705470] env[61768]: DEBUG nova.compute.provider_tree [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.900616] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Releasing lock "refresh_cache-7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.900865] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 581.901062] env[61768]: DEBUG nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.901232] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.928494] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.991606] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Releasing lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.992121] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 581.992340] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 581.992686] env[61768]: DEBUG oslo_concurrency.lockutils [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] Acquired lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.992882] env[61768]: DEBUG nova.network.neutron [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Refreshing network info cache for port 56711858-18ee-4dbf-9461-8f12bb7ab446 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 581.993911] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-294653b7-ab52-4981-8cea-708aa8fa4a6a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.004077] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c53be6-a90c-4737-8f01-af264f3a9fbc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.026247] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e4a7dc76-99a6-4cc2-97c8-f86d72d00d12 could not be found. [ 582.026483] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 582.026665] env[61768]: INFO nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Took 0.03 seconds to destroy the instance on the hypervisor. [ 582.026914] env[61768]: DEBUG oslo.service.loopingcall [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.027149] env[61768]: DEBUG nova.compute.manager [-] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.027316] env[61768]: DEBUG nova.network.neutron [-] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 582.049670] env[61768]: DEBUG nova.network.neutron [-] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.208815] env[61768]: DEBUG nova.scheduler.client.report [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.431650] env[61768]: DEBUG nova.network.neutron [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.526543] env[61768]: DEBUG nova.network.neutron [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.552613] env[61768]: DEBUG nova.network.neutron [-] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.636366] env[61768]: DEBUG nova.network.neutron [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.714123] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.993s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.714462] env[61768]: ERROR nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Traceback (most recent call last): [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self.driver.spawn(context, instance, image_meta, [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] vm_ref = self.build_virtual_machine(instance, [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.714462] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] for vif in network_info: [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return self._sync_wrapper(fn, *args, **kwargs) [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self.wait() [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self[:] = self._gt.wait() [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return self._exit_event.wait() [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] result = hub.switch() [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.714745] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return self.greenlet.switch() [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] result = function(*args, **kwargs) [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] return func(*args, **kwargs) [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] raise e [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] nwinfo = self.network_api.allocate_for_instance( [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] created_port_ids = self._update_ports_for_instance( [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] with excutils.save_and_reraise_exception(): [ 582.715044] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] self.force_reraise() [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] raise self.value [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] updated_port = self._update_port( [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] _ensure_no_port_binding_failure(port) [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] raise exception.PortBindingFailed(port_id=port['id']) [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] nova.exception.PortBindingFailed: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. [ 582.715415] env[61768]: ERROR nova.compute.manager [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] [ 582.715675] env[61768]: DEBUG nova.compute.utils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 582.716512] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.596s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.717960] env[61768]: INFO nova.compute.claims [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 582.720845] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Build of instance 1331a447-7be8-42b8-9d59-6b97742e49d1 was re-scheduled: Binding failed for port f72ba884-e8b4-4f5f-b8b1-d4f475752430, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 582.721298] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 582.721522] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Acquiring lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.721668] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Acquired lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.722049] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.936463] env[61768]: INFO nova.compute.manager [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] [instance: 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7] Took 1.03 seconds to deallocate network for instance. [ 583.010671] env[61768]: DEBUG nova.compute.manager [req-a5eb3ca6-9345-49e4-a638-d0fa5c360fb5 req-66738792-042a-4790-9be8-05e64a933435 service nova] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Received event network-vif-deleted-56711858-18ee-4dbf-9461-8f12bb7ab446 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.056741] env[61768]: INFO nova.compute.manager [-] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Took 1.03 seconds to deallocate network for instance. [ 583.059372] env[61768]: DEBUG nova.compute.claims [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.059525] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.141110] env[61768]: DEBUG oslo_concurrency.lockutils [req-c6d2c7a2-df21-4636-ac81-936757eed6da req-43b7de5f-25ad-47d7-bc5c-0b2b6f05f602 service nova] Releasing lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.248521] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.413323] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.919377] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Releasing lock "refresh_cache-1331a447-7be8-42b8-9d59-6b97742e49d1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.919694] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 583.919801] env[61768]: DEBUG nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.919973] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.942205] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.975276] env[61768]: INFO nova.scheduler.client.report [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Deleted allocations for instance 7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7 [ 584.262104] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f436dd7-b2d3-4e8e-8ec0-ee3691c8751f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.270311] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6120610a-c0f3-4982-a119-e967d926e1dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.307565] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58563482-078d-4633-971d-edc9607c0d91 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.318379] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3439f72-464c-4f6d-b6ae-ace901ee94bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.336711] env[61768]: DEBUG nova.compute.provider_tree [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.446411] env[61768]: DEBUG nova.network.neutron [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.488258] env[61768]: DEBUG oslo_concurrency.lockutils [None req-39f4a436-f101-4366-87b8-377495858b38 tempest-ServersTestFqdnHostnames-1738409658 tempest-ServersTestFqdnHostnames-1738409658-project-member] Lock "7eae5f08-213d-40a1-9d7a-1c4e2b6c3ce7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.814s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.840821] env[61768]: DEBUG nova.scheduler.client.report [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.950699] env[61768]: INFO nova.compute.manager [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] [instance: 1331a447-7be8-42b8-9d59-6b97742e49d1] Took 1.03 seconds to deallocate network for instance. [ 584.993599] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.352167] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.634s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.352167] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.356408] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.699s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.516113] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.858696] env[61768]: DEBUG nova.compute.utils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.860455] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.860618] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 585.935958] env[61768]: DEBUG nova.policy [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fa57679341e41d998188fb2fcc5a444', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '68da18ded44b40f697f9308638a4e3ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.976435] env[61768]: INFO nova.scheduler.client.report [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Deleted allocations for instance 1331a447-7be8-42b8-9d59-6b97742e49d1 [ 586.277207] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Successfully created port: d337de5c-cbd4-4cea-8c96-f3f54122d90b {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.315019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ad1e52-428c-4233-8976-0280545a710c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.321295] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7709fa66-5a71-468f-89fe-20c3ff168ab2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.362781] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ded6653-d568-4f02-bd55-5169f47692c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.366015] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.374176] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c63720c-8bee-43e4-9426-f39cf36d144a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.389269] env[61768]: DEBUG nova.compute.provider_tree [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.489567] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c430cc04-442c-4050-a9ee-0e1d36a26ef0 tempest-ServersV294TestFqdnHostnames-1887362643 tempest-ServersV294TestFqdnHostnames-1887362643-project-member] Lock "1331a447-7be8-42b8-9d59-6b97742e49d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.655s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.892286] env[61768]: DEBUG nova.scheduler.client.report [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.991141] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 587.308653] env[61768]: DEBUG nova.compute.manager [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Received event network-changed-d337de5c-cbd4-4cea-8c96-f3f54122d90b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.309136] env[61768]: DEBUG nova.compute.manager [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Refreshing instance network info cache due to event network-changed-d337de5c-cbd4-4cea-8c96-f3f54122d90b. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 587.309136] env[61768]: DEBUG oslo_concurrency.lockutils [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] Acquiring lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.309374] env[61768]: DEBUG oslo_concurrency.lockutils [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] Acquired lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.309433] env[61768]: DEBUG nova.network.neutron [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Refreshing network info cache for port d337de5c-cbd4-4cea-8c96-f3f54122d90b {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 587.377488] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.381317] env[61768]: ERROR nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. [ 587.381317] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 587.381317] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.381317] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 587.381317] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.381317] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 587.381317] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.381317] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 587.381317] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.381317] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 587.381317] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.381317] env[61768]: ERROR nova.compute.manager raise self.value [ 587.381317] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.381317] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 587.381317] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.381317] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 587.381684] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.381684] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 587.381684] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. [ 587.381684] env[61768]: ERROR nova.compute.manager [ 587.381684] env[61768]: Traceback (most recent call last): [ 587.381684] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 587.381684] env[61768]: listener.cb(fileno) [ 587.381684] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.381684] env[61768]: result = function(*args, **kwargs) [ 587.381684] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.381684] env[61768]: return func(*args, **kwargs) [ 587.381684] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.381684] env[61768]: raise e [ 587.381684] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.381684] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 587.381684] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.381684] env[61768]: created_port_ids = self._update_ports_for_instance( [ 587.381684] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.381684] env[61768]: with excutils.save_and_reraise_exception(): [ 587.381684] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.381684] env[61768]: self.force_reraise() [ 587.381684] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.381684] env[61768]: raise self.value [ 587.381684] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.381684] env[61768]: updated_port = self._update_port( [ 587.381684] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.381684] env[61768]: _ensure_no_port_binding_failure(port) [ 587.381684] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.381684] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 587.382311] env[61768]: nova.exception.PortBindingFailed: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. [ 587.382311] env[61768]: Removing descriptor: 17 [ 587.407217] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.048s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.407217] env[61768]: ERROR nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. [ 587.407217] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Traceback (most recent call last): [ 587.407217] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.407217] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self.driver.spawn(context, instance, image_meta, [ 587.407217] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 587.407217] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.407217] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.407217] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] vm_ref = self.build_virtual_machine(instance, [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] for vif in network_info: [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return self._sync_wrapper(fn, *args, **kwargs) [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self.wait() [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self[:] = self._gt.wait() [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return self._exit_event.wait() [ 587.407562] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] result = hub.switch() [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return self.greenlet.switch() [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] result = function(*args, **kwargs) [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] return func(*args, **kwargs) [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] raise e [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] nwinfo = self.network_api.allocate_for_instance( [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.407894] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] created_port_ids = self._update_ports_for_instance( [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] with excutils.save_and_reraise_exception(): [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] self.force_reraise() [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] raise self.value [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] updated_port = self._update_port( [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] _ensure_no_port_binding_failure(port) [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.408243] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] raise exception.PortBindingFailed(port_id=port['id']) [ 587.408549] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] nova.exception.PortBindingFailed: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. [ 587.408549] env[61768]: ERROR nova.compute.manager [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] [ 587.408549] env[61768]: DEBUG nova.compute.utils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 587.408549] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.618s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.411036] env[61768]: INFO nova.compute.claims [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.416311] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.418016] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.418016] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.418016] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.418016] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.418016] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.418312] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.418312] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.418312] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.418312] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.418571] env[61768]: DEBUG nova.virt.hardware [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.419271] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Build of instance a1b9b618-91b5-429b-8581-9f819e0d81ee was re-scheduled: Binding failed for port 2c5c6dd6-4a1e-4b2c-a3b0-f182007aee55, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 587.419936] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 587.420259] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Acquiring lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.420494] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Acquired lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.420730] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 587.422443] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60dc3c05-6c5b-4077-8503-08e77f629f0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.434738] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623264ca-f232-46fd-bda4-b82e903e4a10 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.452994] env[61768]: ERROR nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Traceback (most recent call last): [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] yield resources [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self.driver.spawn(context, instance, image_meta, [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] vm_ref = self.build_virtual_machine(instance, [ 587.452994] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] for vif in network_info: [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] return self._sync_wrapper(fn, *args, **kwargs) [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self.wait() [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self[:] = self._gt.wait() [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] return self._exit_event.wait() [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 587.453376] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] current.throw(*self._exc) [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] result = function(*args, **kwargs) [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] return func(*args, **kwargs) [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] raise e [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] nwinfo = self.network_api.allocate_for_instance( [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] created_port_ids = self._update_ports_for_instance( [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] with excutils.save_and_reraise_exception(): [ 587.453709] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self.force_reraise() [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] raise self.value [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] updated_port = self._update_port( [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] _ensure_no_port_binding_failure(port) [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] raise exception.PortBindingFailed(port_id=port['id']) [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] nova.exception.PortBindingFailed: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. [ 587.454051] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] [ 587.457027] env[61768]: INFO nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Terminating instance [ 587.458959] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Acquiring lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.528925] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.831033] env[61768]: DEBUG nova.network.neutron [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.959132] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.012950] env[61768]: DEBUG nova.network.neutron [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.108152] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.516806] env[61768]: DEBUG oslo_concurrency.lockutils [req-4bb490d2-e880-45ac-b68b-208c7305d717 req-0be5383d-beec-444d-98b1-0b3ec674914c service nova] Releasing lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.521254] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Acquired lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.521601] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 588.619645] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Releasing lock "refresh_cache-a1b9b618-91b5-429b-8581-9f819e0d81ee" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.619890] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 588.620122] env[61768]: DEBUG nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.620306] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 588.637128] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.889393] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40cd0b95-7635-4f90-affa-79d10e8298b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.897059] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7d3a3e-e39c-4233-a12a-8a5dc45d1cdb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.930738] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd17a95-f2b3-482c-b349-9fae9a1be2a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.938652] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8152c784-587e-424e-a7c0-04314f3703da {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.954631] env[61768]: DEBUG nova.compute.provider_tree [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.053248] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.139815] env[61768]: DEBUG nova.network.neutron [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.240663] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 589.349920] env[61768]: DEBUG nova.compute.manager [req-bc87e454-faf8-4e07-8757-92aae7e99705 req-e63be1d8-0ae6-4660-94fb-6632e4004075 service nova] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Received event network-vif-deleted-d337de5c-cbd4-4cea-8c96-f3f54122d90b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.458074] env[61768]: DEBUG nova.scheduler.client.report [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.647143] env[61768]: INFO nova.compute.manager [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] [instance: a1b9b618-91b5-429b-8581-9f819e0d81ee] Took 1.02 seconds to deallocate network for instance. [ 589.745141] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Releasing lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.745141] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.745141] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 589.745141] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b8743349-5f07-4b02-b2bc-ab3105dac2b8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.755951] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ce48b0-ad08-4311-a430-63b3bb6c86e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.780184] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0103eded-5a0b-4b7f-b98f-63d7866d501c could not be found. [ 589.780184] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 589.780184] env[61768]: INFO nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 589.780184] env[61768]: DEBUG oslo.service.loopingcall [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.780184] env[61768]: DEBUG nova.compute.manager [-] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.780184] env[61768]: DEBUG nova.network.neutron [-] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 589.797296] env[61768]: DEBUG nova.network.neutron [-] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 589.970022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.970022] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.970733] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.998s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.972714] env[61768]: INFO nova.compute.claims [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.302033] env[61768]: DEBUG nova.network.neutron [-] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.481583] env[61768]: DEBUG nova.compute.utils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.487008] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.487008] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 590.558153] env[61768]: DEBUG nova.policy [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b644f2c92abd4559a88d99199314268f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ae392a1509054e6ab7e8698b20db16ff', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.694180] env[61768]: INFO nova.scheduler.client.report [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Deleted allocations for instance a1b9b618-91b5-429b-8581-9f819e0d81ee [ 590.806318] env[61768]: INFO nova.compute.manager [-] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Took 1.03 seconds to deallocate network for instance. [ 590.809649] env[61768]: DEBUG nova.compute.claims [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.809649] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.948514] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Successfully created port: 6cebe1cf-0a80-4fd5-8dac-209ebea16918 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 590.985219] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.716631] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4756209-0487-435f-bc28-34fadadac4c7 tempest-ServerMetadataNegativeTestJSON-148337351 tempest-ServerMetadataNegativeTestJSON-148337351-project-member] Lock "a1b9b618-91b5-429b-8581-9f819e0d81ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.446s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.104052] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248eb717-83eb-4b45-b4ce-a0fcbc5a0c07 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.114112] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d690cbb-bbb7-48db-b49b-89272b78abef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.145049] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7647a0a2-d87b-4a99-83e7-608b8ddbb1f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.152692] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719f36b4-8ebf-4da8-8f35-9e586c62578d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.171556] env[61768]: DEBUG nova.compute.provider_tree [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.222657] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 592.228346] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 592.258956] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 592.259270] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 592.260177] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 592.260425] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 592.261154] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 592.261154] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 592.262119] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 592.262119] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 592.262119] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 592.262119] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 592.262665] env[61768]: DEBUG nova.virt.hardware [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 592.264872] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e9c1e5-3eb6-49dc-a15d-9d5f3fc2ab18 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.274945] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c99bfc7-d623-4f0d-91ef-9e31cc769265 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.351921] env[61768]: ERROR nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. [ 592.351921] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.351921] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.351921] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.351921] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.351921] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.351921] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.351921] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.351921] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.351921] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 592.351921] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.351921] env[61768]: ERROR nova.compute.manager raise self.value [ 592.351921] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.351921] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.351921] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.351921] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.352324] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.352324] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.352324] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. [ 592.352324] env[61768]: ERROR nova.compute.manager [ 592.352324] env[61768]: Traceback (most recent call last): [ 592.352324] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.352324] env[61768]: listener.cb(fileno) [ 592.352324] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.352324] env[61768]: result = function(*args, **kwargs) [ 592.352324] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.352324] env[61768]: return func(*args, **kwargs) [ 592.352324] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.352324] env[61768]: raise e [ 592.352324] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.352324] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 592.352324] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.352324] env[61768]: created_port_ids = self._update_ports_for_instance( [ 592.352324] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.352324] env[61768]: with excutils.save_and_reraise_exception(): [ 592.352324] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.352324] env[61768]: self.force_reraise() [ 592.352324] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.352324] env[61768]: raise self.value [ 592.352324] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.352324] env[61768]: updated_port = self._update_port( [ 592.352324] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.352324] env[61768]: _ensure_no_port_binding_failure(port) [ 592.352324] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.352324] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.353134] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. [ 592.353134] env[61768]: Removing descriptor: 14 [ 592.353134] env[61768]: ERROR nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Traceback (most recent call last): [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] yield resources [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self.driver.spawn(context, instance, image_meta, [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.353134] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] vm_ref = self.build_virtual_machine(instance, [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] for vif in network_info: [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return self._sync_wrapper(fn, *args, **kwargs) [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self.wait() [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self[:] = self._gt.wait() [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return self._exit_event.wait() [ 592.353433] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] result = hub.switch() [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return self.greenlet.switch() [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] result = function(*args, **kwargs) [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return func(*args, **kwargs) [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] raise e [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] nwinfo = self.network_api.allocate_for_instance( [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 592.353735] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] created_port_ids = self._update_ports_for_instance( [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] with excutils.save_and_reraise_exception(): [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self.force_reraise() [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] raise self.value [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] updated_port = self._update_port( [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] _ensure_no_port_binding_failure(port) [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.354042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] raise exception.PortBindingFailed(port_id=port['id']) [ 592.354324] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] nova.exception.PortBindingFailed: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. [ 592.354324] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] [ 592.354324] env[61768]: INFO nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Terminating instance [ 592.355695] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.355853] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquired lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.356168] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 592.644209] env[61768]: DEBUG nova.compute.manager [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Received event network-changed-6cebe1cf-0a80-4fd5-8dac-209ebea16918 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.644430] env[61768]: DEBUG nova.compute.manager [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Refreshing instance network info cache due to event network-changed-6cebe1cf-0a80-4fd5-8dac-209ebea16918. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.644634] env[61768]: DEBUG oslo_concurrency.lockutils [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] Acquiring lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.674269] env[61768]: DEBUG nova.scheduler.client.report [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.750182] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.889128] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 592.926625] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "3401be45-c3ab-4780-81a7-a93b05742414" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.927413] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "3401be45-c3ab-4780-81a7-a93b05742414" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.994469] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.185288] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.212s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.185288] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.187519] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.083s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.190660] env[61768]: INFO nova.compute.claims [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.500587] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Releasing lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.500587] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.500587] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 593.500587] env[61768]: DEBUG oslo_concurrency.lockutils [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] Acquired lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 593.500587] env[61768]: DEBUG nova.network.neutron [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Refreshing network info cache for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 593.500752] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6de4a2a7-0540-4c2b-afd1-52353e4afdd2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.513762] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d13c17-2938-45b2-a0c3-6ed71574e20a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.537996] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 70e6d944-f978-4d5b-b396-e0e913db0485 could not be found. [ 593.538277] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 593.538858] env[61768]: INFO nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Took 0.04 seconds to destroy the instance on the hypervisor. [ 593.538858] env[61768]: DEBUG oslo.service.loopingcall [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.538988] env[61768]: DEBUG nova.compute.manager [-] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.538988] env[61768]: DEBUG nova.network.neutron [-] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 593.567896] env[61768]: DEBUG nova.network.neutron [-] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 593.696812] env[61768]: DEBUG nova.compute.utils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.698257] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.698424] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.780154] env[61768]: DEBUG nova.policy [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdf18ac0ee9b45db8a2bcef76a0b35dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a97ab775489436da69dc4a3dc1ff7ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.034014] env[61768]: DEBUG nova.network.neutron [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.070206] env[61768]: DEBUG nova.network.neutron [-] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.149443] env[61768]: DEBUG nova.network.neutron [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.209081] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 594.316858] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Successfully created port: b43c8943-f264-45d4-8fcc-80670ee662e8 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.573284] env[61768]: INFO nova.compute.manager [-] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Took 1.03 seconds to deallocate network for instance. [ 594.575678] env[61768]: DEBUG nova.compute.claims [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.575854] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.655591] env[61768]: DEBUG oslo_concurrency.lockutils [req-4b8bdd74-afbf-4826-ad41-0ce26a126f3c req-03dcbe3a-69cb-447a-8953-aa2e3cfa10f5 service nova] Releasing lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 594.691383] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e471b01b-b665-436b-add3-ae7a1bda5f05 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.701392] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b56c293d-85e7-49e3-b3f8-a3472cb92e7b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.739697] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba6b632-a478-4a85-afa5-36f37f91ecd2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.750910] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7b95431-216e-49f3-8f4a-824a4c5e2059 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.767745] env[61768]: DEBUG nova.compute.provider_tree [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.911962] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.912660] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.118296] env[61768]: DEBUG nova.compute.manager [req-0b8b44c0-0c7a-4d51-ba31-f8e1a71774fb req-97752526-7524-4eb7-b2cd-105eb791b604 service nova] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Received event network-vif-deleted-6cebe1cf-0a80-4fd5-8dac-209ebea16918 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.244026] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.264293] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.264604] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.264917] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.265358] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.265555] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.265713] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.266132] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.266490] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.266676] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.266850] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.267200] env[61768]: DEBUG nova.virt.hardware [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.268320] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b0c5f6-22cb-4063-a625-6a03cfba5eda {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.273007] env[61768]: DEBUG nova.scheduler.client.report [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.282913] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d334ebce-3d67-4c8e-b268-c8e11f45b2b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.552066] env[61768]: ERROR nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. [ 595.552066] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.552066] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.552066] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.552066] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.552066] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.552066] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.552066] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.552066] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.552066] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 595.552066] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.552066] env[61768]: ERROR nova.compute.manager raise self.value [ 595.552066] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.552066] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.552066] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.552066] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.552572] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.552572] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.552572] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. [ 595.552572] env[61768]: ERROR nova.compute.manager [ 595.552572] env[61768]: Traceback (most recent call last): [ 595.552572] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.552572] env[61768]: listener.cb(fileno) [ 595.552572] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.552572] env[61768]: result = function(*args, **kwargs) [ 595.552572] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.552572] env[61768]: return func(*args, **kwargs) [ 595.552572] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.552572] env[61768]: raise e [ 595.552572] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.552572] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 595.552572] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.552572] env[61768]: created_port_ids = self._update_ports_for_instance( [ 595.552572] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.552572] env[61768]: with excutils.save_and_reraise_exception(): [ 595.552572] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.552572] env[61768]: self.force_reraise() [ 595.552572] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.552572] env[61768]: raise self.value [ 595.552572] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.552572] env[61768]: updated_port = self._update_port( [ 595.552572] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.552572] env[61768]: _ensure_no_port_binding_failure(port) [ 595.552572] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.552572] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.553375] env[61768]: nova.exception.PortBindingFailed: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. [ 595.553375] env[61768]: Removing descriptor: 14 [ 595.553375] env[61768]: ERROR nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Traceback (most recent call last): [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] yield resources [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self.driver.spawn(context, instance, image_meta, [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.553375] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] vm_ref = self.build_virtual_machine(instance, [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] for vif in network_info: [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return self._sync_wrapper(fn, *args, **kwargs) [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self.wait() [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self[:] = self._gt.wait() [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return self._exit_event.wait() [ 595.553668] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] result = hub.switch() [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return self.greenlet.switch() [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] result = function(*args, **kwargs) [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return func(*args, **kwargs) [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] raise e [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] nwinfo = self.network_api.allocate_for_instance( [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.553966] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] created_port_ids = self._update_ports_for_instance( [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] with excutils.save_and_reraise_exception(): [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self.force_reraise() [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] raise self.value [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] updated_port = self._update_port( [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] _ensure_no_port_binding_failure(port) [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.554286] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] raise exception.PortBindingFailed(port_id=port['id']) [ 595.554586] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] nova.exception.PortBindingFailed: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. [ 595.554586] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] [ 595.554586] env[61768]: INFO nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Terminating instance [ 595.556770] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.557009] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquired lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.557576] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.778245] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.591s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.781101] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.784021] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.619s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.080021] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.139416] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.289496] env[61768]: DEBUG nova.compute.utils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.293866] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.295106] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 596.347810] env[61768]: DEBUG nova.policy [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '353dee3c759743d4867459eb76913691', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42e166d62eb94f2bbe4de986d8479777', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.641755] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Releasing lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.642185] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.642386] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 596.642677] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-baec70ce-a735-4d0f-a17a-c15a76b28758 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.658547] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd01765-0374-48b7-a4f9-5ed6aa94400b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.686722] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fcddfd64-74bf-4a05-9799-0b7fee7416b2 could not be found. [ 596.686978] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 596.687181] env[61768]: INFO nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 596.687932] env[61768]: DEBUG oslo.service.loopingcall [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.687932] env[61768]: DEBUG nova.compute.manager [-] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.687932] env[61768]: DEBUG nova.network.neutron [-] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.720362] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b47849-9f72-4eb2-8321-347fa9050e83 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.726018] env[61768]: DEBUG nova.network.neutron [-] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 596.729762] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40aecb04-5da1-42a2-880c-9638893ebfd2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.761357] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90501593-c7cb-4561-ac8c-251a52bdbb24 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.768483] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1185fbf2-31db-41b5-afaf-80fa2fab6447 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.784958] env[61768]: DEBUG nova.compute.provider_tree [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.786555] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Successfully created port: 8dda1026-0a57-4b2a-bbf4-7bd24c46175d {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.795936] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.881769] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 596.882044] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.204509] env[61768]: DEBUG nova.compute.manager [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Received event network-changed-b43c8943-f264-45d4-8fcc-80670ee662e8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.204771] env[61768]: DEBUG nova.compute.manager [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Refreshing instance network info cache due to event network-changed-b43c8943-f264-45d4-8fcc-80670ee662e8. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 597.204917] env[61768]: DEBUG oslo_concurrency.lockutils [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] Acquiring lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.205070] env[61768]: DEBUG oslo_concurrency.lockutils [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] Acquired lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.205611] env[61768]: DEBUG nova.network.neutron [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Refreshing network info cache for port b43c8943-f264-45d4-8fcc-80670ee662e8 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.226064] env[61768]: DEBUG nova.network.neutron [-] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.292104] env[61768]: DEBUG nova.scheduler.client.report [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.388581] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.388752] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 597.388879] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Rebuilding the list of instances to heal {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 597.728742] env[61768]: INFO nova.compute.manager [-] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Took 1.04 seconds to deallocate network for instance. [ 597.734618] env[61768]: DEBUG nova.compute.claims [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.734813] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.739037] env[61768]: DEBUG nova.network.neutron [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.799963] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.799963] env[61768]: ERROR nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. [ 597.799963] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Traceback (most recent call last): [ 597.799963] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 597.799963] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self.driver.spawn(context, instance, image_meta, [ 597.799963] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 597.799963] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 597.799963] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 597.799963] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] vm_ref = self.build_virtual_machine(instance, [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] vif_infos = vmwarevif.get_vif_info(self._session, [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] for vif in network_info: [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return self._sync_wrapper(fn, *args, **kwargs) [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self.wait() [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self[:] = self._gt.wait() [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return self._exit_event.wait() [ 597.800202] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] result = hub.switch() [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return self.greenlet.switch() [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] result = function(*args, **kwargs) [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] return func(*args, **kwargs) [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] raise e [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] nwinfo = self.network_api.allocate_for_instance( [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 597.800492] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] created_port_ids = self._update_ports_for_instance( [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] with excutils.save_and_reraise_exception(): [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] self.force_reraise() [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] raise self.value [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] updated_port = self._update_port( [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] _ensure_no_port_binding_failure(port) [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 597.800784] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] raise exception.PortBindingFailed(port_id=port['id']) [ 597.801062] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] nova.exception.PortBindingFailed: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. [ 597.801062] env[61768]: ERROR nova.compute.manager [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] [ 597.801062] env[61768]: DEBUG nova.compute.utils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 597.801062] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.962s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.805418] env[61768]: INFO nova.compute.claims [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.812923] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Build of instance 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe was re-scheduled: Binding failed for port dfbb10fa-71a5-4f65-beff-de178b6f1680, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 597.813057] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 597.813281] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.813396] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.813551] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 597.818059] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.859885] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.860216] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.860384] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.861026] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.861026] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.861026] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.861208] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.861384] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.861553] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.861762] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.861896] env[61768]: DEBUG nova.virt.hardware [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.863435] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a48d8a1-f16e-4cee-905a-c801c6383fcb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.866349] env[61768]: DEBUG nova.network.neutron [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.879436] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c0726f0-8503-448f-aae8-752d5f2d7abc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.897945] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.898352] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.898352] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.898352] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.898520] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 597.898594] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Didn't find any instances for network info cache update. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 597.900958] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.901276] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.901752] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.901923] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.902085] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.902238] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 597.902368] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 597.902514] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 598.147995] env[61768]: ERROR nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. [ 598.147995] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.147995] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.147995] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.147995] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.147995] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.147995] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.147995] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.147995] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.147995] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 598.147995] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.147995] env[61768]: ERROR nova.compute.manager raise self.value [ 598.147995] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.147995] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.147995] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.147995] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.148561] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.148561] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.148561] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. [ 598.148561] env[61768]: ERROR nova.compute.manager [ 598.148561] env[61768]: Traceback (most recent call last): [ 598.148561] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.148561] env[61768]: listener.cb(fileno) [ 598.148561] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.148561] env[61768]: result = function(*args, **kwargs) [ 598.148561] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.148561] env[61768]: return func(*args, **kwargs) [ 598.148561] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.148561] env[61768]: raise e [ 598.148561] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.148561] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 598.148561] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.148561] env[61768]: created_port_ids = self._update_ports_for_instance( [ 598.148561] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.148561] env[61768]: with excutils.save_and_reraise_exception(): [ 598.148561] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.148561] env[61768]: self.force_reraise() [ 598.148561] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.148561] env[61768]: raise self.value [ 598.148561] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.148561] env[61768]: updated_port = self._update_port( [ 598.148561] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.148561] env[61768]: _ensure_no_port_binding_failure(port) [ 598.148561] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.148561] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.149347] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. [ 598.149347] env[61768]: Removing descriptor: 14 [ 598.149347] env[61768]: ERROR nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Traceback (most recent call last): [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] yield resources [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self.driver.spawn(context, instance, image_meta, [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.149347] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] vm_ref = self.build_virtual_machine(instance, [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] for vif in network_info: [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return self._sync_wrapper(fn, *args, **kwargs) [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self.wait() [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self[:] = self._gt.wait() [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return self._exit_event.wait() [ 598.149679] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] result = hub.switch() [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return self.greenlet.switch() [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] result = function(*args, **kwargs) [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return func(*args, **kwargs) [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] raise e [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] nwinfo = self.network_api.allocate_for_instance( [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.150040] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] created_port_ids = self._update_ports_for_instance( [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] with excutils.save_and_reraise_exception(): [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self.force_reraise() [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] raise self.value [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] updated_port = self._update_port( [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] _ensure_no_port_binding_failure(port) [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.150418] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] raise exception.PortBindingFailed(port_id=port['id']) [ 598.150736] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] nova.exception.PortBindingFailed: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. [ 598.150736] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] [ 598.150736] env[61768]: INFO nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Terminating instance [ 598.150736] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Acquiring lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.150736] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Acquired lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.150883] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.338522] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.368953] env[61768]: DEBUG oslo_concurrency.lockutils [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] Releasing lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.369195] env[61768]: DEBUG nova.compute.manager [req-7d22fe36-6288-47ae-9768-aa34b88ff578 req-4a95ed2b-7661-4797-aaa7-3e4c24bf7c32 service nova] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Received event network-vif-deleted-b43c8943-f264-45d4-8fcc-80670ee662e8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.405979] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.418584] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.667188] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.749073] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.920866] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.921127] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 598.921318] env[61768]: DEBUG nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.921792] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 598.939016] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.212942] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7aa7b1-3ba6-4d02-acb4-70ec6c5d1050 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.221685] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc70e30-b9d1-42de-b513-1cada0f7aa9e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.250906] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Releasing lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.251343] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.251714] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.251968] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03dc7ebc-4677-49d6-9123-3b190651cf3a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.254233] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf51589-e873-4785-b204-2718f34619fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.259029] env[61768]: DEBUG nova.compute.manager [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Received event network-changed-8dda1026-0a57-4b2a-bbf4-7bd24c46175d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.259029] env[61768]: DEBUG nova.compute.manager [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Refreshing instance network info cache due to event network-changed-8dda1026-0a57-4b2a-bbf4-7bd24c46175d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 599.259029] env[61768]: DEBUG oslo_concurrency.lockutils [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] Acquiring lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.259029] env[61768]: DEBUG oslo_concurrency.lockutils [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] Acquired lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.259029] env[61768]: DEBUG nova.network.neutron [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Refreshing network info cache for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 599.266166] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2eff5a4-f279-423b-9220-85c1ef503d95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.273197] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d0656b-f5b7-4f1c-b9b8-376954fc9278 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.293554] env[61768]: DEBUG nova.compute.provider_tree [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.298503] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f0be276a-ddad-4ae0-a232-d0fb8463a3bc could not be found. [ 599.298503] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.298568] env[61768]: INFO nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 599.298806] env[61768]: DEBUG oslo.service.loopingcall [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.299212] env[61768]: DEBUG nova.compute.manager [-] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.299307] env[61768]: DEBUG nova.network.neutron [-] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.314323] env[61768]: DEBUG nova.network.neutron [-] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.444462] env[61768]: DEBUG nova.network.neutron [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.775831] env[61768]: DEBUG nova.network.neutron [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.799978] env[61768]: DEBUG nova.scheduler.client.report [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 599.815858] env[61768]: DEBUG nova.network.neutron [-] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.844398] env[61768]: DEBUG nova.network.neutron [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.949305] env[61768]: INFO nova.compute.manager [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe] Took 1.03 seconds to deallocate network for instance. [ 600.304477] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.306528] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.310104] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.250s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.318193] env[61768]: INFO nova.compute.manager [-] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Took 1.02 seconds to deallocate network for instance. [ 600.320082] env[61768]: DEBUG nova.compute.claims [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.320257] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.346890] env[61768]: DEBUG oslo_concurrency.lockutils [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] Releasing lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.347050] env[61768]: DEBUG nova.compute.manager [req-3afa6e3d-5f62-4398-bfdf-d77b1aa16155 req-59d371ff-0a6e-4dce-8472-6fa0407d91fd service nova] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Received event network-vif-deleted-8dda1026-0a57-4b2a-bbf4-7bd24c46175d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.814835] env[61768]: DEBUG nova.compute.utils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.820221] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.821238] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 600.886504] env[61768]: DEBUG nova.policy [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bdf18ac0ee9b45db8a2bcef76a0b35dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0a97ab775489436da69dc4a3dc1ff7ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 600.975820] env[61768]: INFO nova.scheduler.client.report [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted allocations for instance 7a8228e2-3ab3-4b87-ad38-4246e33f0dfe [ 601.203213] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Successfully created port: de30f4a1-b25d-44ab-a19e-8685a941cc60 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.221871] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6a5c0e0-954c-4ea6-8145-a017fe0bac86 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.230227] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994e4676-e658-4190-804e-458069ad62a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.263425] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113991b0-f97d-408a-90eb-f7ea164e4486 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.271148] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38cdf7f9-4513-4b53-b633-e2cc455e7c69 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.284578] env[61768]: DEBUG nova.compute.provider_tree [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 601.320731] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.486126] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56cd553e-f8ec-4f5c-9885-042915f92f00 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "7a8228e2-3ab3-4b87-ad38-4246e33f0dfe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.290s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.790737] env[61768]: DEBUG nova.scheduler.client.report [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.990155] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.082103] env[61768]: DEBUG nova.compute.manager [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Received event network-changed-de30f4a1-b25d-44ab-a19e-8685a941cc60 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.082402] env[61768]: DEBUG nova.compute.manager [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Refreshing instance network info cache due to event network-changed-de30f4a1-b25d-44ab-a19e-8685a941cc60. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 602.082638] env[61768]: DEBUG oslo_concurrency.lockutils [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] Acquiring lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.082767] env[61768]: DEBUG oslo_concurrency.lockutils [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] Acquired lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.082930] env[61768]: DEBUG nova.network.neutron [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Refreshing network info cache for port de30f4a1-b25d-44ab-a19e-8685a941cc60 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 602.226292] env[61768]: ERROR nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. [ 602.226292] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.226292] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.226292] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.226292] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.226292] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.226292] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.226292] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.226292] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.226292] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 602.226292] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.226292] env[61768]: ERROR nova.compute.manager raise self.value [ 602.226292] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.226292] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.226292] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.226292] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.226737] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.226737] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.226737] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. [ 602.226737] env[61768]: ERROR nova.compute.manager [ 602.226737] env[61768]: Traceback (most recent call last): [ 602.226737] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.226737] env[61768]: listener.cb(fileno) [ 602.226737] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.226737] env[61768]: result = function(*args, **kwargs) [ 602.226737] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.226737] env[61768]: return func(*args, **kwargs) [ 602.226737] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.226737] env[61768]: raise e [ 602.226737] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.226737] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 602.226737] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.226737] env[61768]: created_port_ids = self._update_ports_for_instance( [ 602.226737] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.226737] env[61768]: with excutils.save_and_reraise_exception(): [ 602.226737] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.226737] env[61768]: self.force_reraise() [ 602.226737] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.226737] env[61768]: raise self.value [ 602.226737] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.226737] env[61768]: updated_port = self._update_port( [ 602.226737] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.226737] env[61768]: _ensure_no_port_binding_failure(port) [ 602.226737] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.226737] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.227553] env[61768]: nova.exception.PortBindingFailed: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. [ 602.227553] env[61768]: Removing descriptor: 14 [ 602.296724] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.987s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.297322] env[61768]: ERROR nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Traceback (most recent call last): [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self.driver.spawn(context, instance, image_meta, [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] vm_ref = self.build_virtual_machine(instance, [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.297322] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] for vif in network_info: [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return self._sync_wrapper(fn, *args, **kwargs) [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self.wait() [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self[:] = self._gt.wait() [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return self._exit_event.wait() [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] result = hub.switch() [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.297754] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return self.greenlet.switch() [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] result = function(*args, **kwargs) [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] return func(*args, **kwargs) [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] raise e [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] nwinfo = self.network_api.allocate_for_instance( [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] created_port_ids = self._update_ports_for_instance( [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] with excutils.save_and_reraise_exception(): [ 602.299717] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] self.force_reraise() [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] raise self.value [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] updated_port = self._update_port( [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] _ensure_no_port_binding_failure(port) [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] raise exception.PortBindingFailed(port_id=port['id']) [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] nova.exception.PortBindingFailed: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. [ 602.300138] env[61768]: ERROR nova.compute.manager [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] [ 602.300454] env[61768]: DEBUG nova.compute.utils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 602.300454] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.783s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.303710] env[61768]: INFO nova.compute.claims [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 602.304129] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Build of instance e4a7dc76-99a6-4cc2-97c8-f86d72d00d12 was re-scheduled: Binding failed for port 56711858-18ee-4dbf-9461-8f12bb7ab446, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 602.305043] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 602.305043] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Acquiring lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.305043] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Acquired lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.305252] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.330639] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.359936] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.360240] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.360407] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.360591] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.360739] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.360887] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.362899] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.363166] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.364523] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.364820] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.365121] env[61768]: DEBUG nova.virt.hardware [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.366506] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e983ac92-dd78-4f92-bb1c-16c9e0d0a78d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.377457] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e784d2e-ef18-4f4f-917d-6221ae16e9c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.392644] env[61768]: ERROR nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Traceback (most recent call last): [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] yield resources [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self.driver.spawn(context, instance, image_meta, [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] vm_ref = self.build_virtual_machine(instance, [ 602.392644] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] for vif in network_info: [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] return self._sync_wrapper(fn, *args, **kwargs) [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self.wait() [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self[:] = self._gt.wait() [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] return self._exit_event.wait() [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 602.392994] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] current.throw(*self._exc) [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] result = function(*args, **kwargs) [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] return func(*args, **kwargs) [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] raise e [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] nwinfo = self.network_api.allocate_for_instance( [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] created_port_ids = self._update_ports_for_instance( [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] with excutils.save_and_reraise_exception(): [ 602.393373] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self.force_reraise() [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] raise self.value [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] updated_port = self._update_port( [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] _ensure_no_port_binding_failure(port) [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] raise exception.PortBindingFailed(port_id=port['id']) [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] nova.exception.PortBindingFailed: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. [ 602.393728] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] [ 602.393728] env[61768]: INFO nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Terminating instance [ 602.395032] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.508776] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.604159] env[61768]: DEBUG nova.network.neutron [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.668931] env[61768]: DEBUG nova.network.neutron [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.825154] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.836537] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "efbde802-8cb0-4563-a776-3722a3889afe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.837198] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "efbde802-8cb0-4563-a776-3722a3889afe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.918435] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.171836] env[61768]: DEBUG oslo_concurrency.lockutils [req-76b546a1-c36e-42f9-a5c1-0c11d10c8bef req-33649359-65c0-425c-83f3-e6eba5304390 service nova] Releasing lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.172282] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquired lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.172508] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 603.421573] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Releasing lock "refresh_cache-e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.421831] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 603.422052] env[61768]: DEBUG nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.422203] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.440886] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.699672] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.704432] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b677fc0b-79cc-4d9b-8239-cccbf7114b0c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.711687] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f011cce1-adbd-4022-8dae-c0200b9cfe5a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.743533] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103b88ca-13f0-4a1e-b0d2-b38854852f33 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.751185] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9999eb5-d6a7-444c-a30b-4f2afffff6f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.765455] env[61768]: DEBUG nova.compute.provider_tree [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.767239] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.945563] env[61768]: DEBUG nova.network.neutron [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.106729] env[61768]: DEBUG nova.compute.manager [req-ea408161-eb68-4f64-8946-50f00a2b3584 req-849e1723-0c48-45cb-bc8b-57edbf056bf0 service nova] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Received event network-vif-deleted-de30f4a1-b25d-44ab-a19e-8685a941cc60 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 604.269486] env[61768]: DEBUG nova.scheduler.client.report [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.273138] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Releasing lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 604.273580] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 604.273784] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 604.274072] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f78286f7-161f-434c-841e-39cd4c45a26e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.283239] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295930ed-aedf-4bfa-af8f-435803c56abe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.305605] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aadc35e7-7aea-4f86-be3a-9b5322f350a5 could not be found. [ 604.305845] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 604.306064] env[61768]: INFO nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 604.306321] env[61768]: DEBUG oslo.service.loopingcall [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.306537] env[61768]: DEBUG nova.compute.manager [-] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.306633] env[61768]: DEBUG nova.network.neutron [-] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 604.329304] env[61768]: DEBUG nova.network.neutron [-] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.449186] env[61768]: INFO nova.compute.manager [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] [instance: e4a7dc76-99a6-4cc2-97c8-f86d72d00d12] Took 1.03 seconds to deallocate network for instance. [ 604.778846] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.779422] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.782365] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.254s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.784020] env[61768]: INFO nova.compute.claims [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.831696] env[61768]: DEBUG nova.network.neutron [-] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.288036] env[61768]: DEBUG nova.compute.utils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 605.291465] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 605.291639] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 605.330187] env[61768]: DEBUG nova.policy [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bec6271df93e432a941fc02a925de97b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a1f1f09a7af14f3e8d184147dfb06c98', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 605.334135] env[61768]: INFO nova.compute.manager [-] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Took 1.03 seconds to deallocate network for instance. [ 605.336424] env[61768]: DEBUG nova.compute.claims [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 605.336601] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.495866] env[61768]: INFO nova.scheduler.client.report [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Deleted allocations for instance e4a7dc76-99a6-4cc2-97c8-f86d72d00d12 [ 605.670318] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Successfully created port: 90da7add-917d-486c-b346-1e3eb5b991a7 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.792737] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 606.006516] env[61768]: DEBUG oslo_concurrency.lockutils [None req-76a13b64-e8f0-41ea-83d7-8bb7c4dc453f tempest-ServerActionsTestOtherA-1398658537 tempest-ServerActionsTestOtherA-1398658537-project-member] Lock "e4a7dc76-99a6-4cc2-97c8-f86d72d00d12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.082s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.249686] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7e3e2df-1bf0-4f2d-8bf0-90113dac2df7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.257653] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0261bb-1c52-4f0f-8a3e-4212866589b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.291100] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c817fd-4206-4e0b-9e51-3a5990323e87 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.295794] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-323b1af1-0712-4489-b2c7-8ba311952f79 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.312195] env[61768]: DEBUG nova.compute.provider_tree [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.514657] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.665636] env[61768]: DEBUG nova.compute.manager [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Received event network-changed-90da7add-917d-486c-b346-1e3eb5b991a7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.666030] env[61768]: DEBUG nova.compute.manager [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Refreshing instance network info cache due to event network-changed-90da7add-917d-486c-b346-1e3eb5b991a7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 606.666135] env[61768]: DEBUG oslo_concurrency.lockutils [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] Acquiring lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.666317] env[61768]: DEBUG oslo_concurrency.lockutils [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] Acquired lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.666711] env[61768]: DEBUG nova.network.neutron [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Refreshing network info cache for port 90da7add-917d-486c-b346-1e3eb5b991a7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 606.808304] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.818018] env[61768]: DEBUG nova.scheduler.client.report [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.836186] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.839577] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.839577] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.839577] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.839577] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.839577] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.839774] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.839774] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.839774] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.839774] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.839774] env[61768]: DEBUG nova.virt.hardware [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.839911] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e74f17-57fa-4e10-8e18-8b5b52b70822 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.849422] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adb6920-2752-4ff3-a2b0-24d67c6bacb3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.036555] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 607.099534] env[61768]: ERROR nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. [ 607.099534] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 607.099534] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.099534] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 607.099534] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.099534] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 607.099534] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.099534] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 607.099534] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.099534] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 607.099534] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.099534] env[61768]: ERROR nova.compute.manager raise self.value [ 607.099534] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.099534] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 607.099534] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.099534] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 607.100287] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.100287] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 607.100287] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. [ 607.100287] env[61768]: ERROR nova.compute.manager [ 607.100287] env[61768]: Traceback (most recent call last): [ 607.100287] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 607.100287] env[61768]: listener.cb(fileno) [ 607.100287] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.100287] env[61768]: result = function(*args, **kwargs) [ 607.100287] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.100287] env[61768]: return func(*args, **kwargs) [ 607.100287] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.100287] env[61768]: raise e [ 607.100287] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.100287] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 607.100287] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.100287] env[61768]: created_port_ids = self._update_ports_for_instance( [ 607.100287] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.100287] env[61768]: with excutils.save_and_reraise_exception(): [ 607.100287] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.100287] env[61768]: self.force_reraise() [ 607.100287] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.100287] env[61768]: raise self.value [ 607.100287] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.100287] env[61768]: updated_port = self._update_port( [ 607.100287] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.100287] env[61768]: _ensure_no_port_binding_failure(port) [ 607.100287] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.100287] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 607.101121] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. [ 607.101121] env[61768]: Removing descriptor: 14 [ 607.101121] env[61768]: ERROR nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Traceback (most recent call last): [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] yield resources [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self.driver.spawn(context, instance, image_meta, [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self._vmops.spawn(context, instance, image_meta, injected_files, [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 607.101121] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] vm_ref = self.build_virtual_machine(instance, [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] vif_infos = vmwarevif.get_vif_info(self._session, [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] for vif in network_info: [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return self._sync_wrapper(fn, *args, **kwargs) [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self.wait() [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self[:] = self._gt.wait() [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return self._exit_event.wait() [ 607.101400] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] result = hub.switch() [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return self.greenlet.switch() [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] result = function(*args, **kwargs) [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return func(*args, **kwargs) [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] raise e [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] nwinfo = self.network_api.allocate_for_instance( [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 607.101681] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] created_port_ids = self._update_ports_for_instance( [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] with excutils.save_and_reraise_exception(): [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self.force_reraise() [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] raise self.value [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] updated_port = self._update_port( [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] _ensure_no_port_binding_failure(port) [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 607.101966] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] raise exception.PortBindingFailed(port_id=port['id']) [ 607.102245] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] nova.exception.PortBindingFailed: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. [ 607.102245] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] [ 607.102245] env[61768]: INFO nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Terminating instance [ 607.103818] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Acquiring lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 607.185187] env[61768]: DEBUG nova.network.neutron [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 607.321192] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.539s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.321712] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.324461] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.515s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.345312] env[61768]: DEBUG nova.network.neutron [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.834440] env[61768]: DEBUG nova.compute.utils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.840220] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.840220] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.850548] env[61768]: DEBUG oslo_concurrency.lockutils [req-2c5575c1-522a-45aa-ac7b-227b66283274 req-86e11022-6b48-4e1b-92f2-ab45f9f5e4b1 service nova] Releasing lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.850548] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Acquired lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.850548] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 607.911210] env[61768]: DEBUG nova.policy [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f538a940471d4bc3941415c41d4ef9d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f989ffd7857b42599b4e3585d7f39506', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 608.243361] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Successfully created port: 617cd2e1-7d6a-471a-a06a-973c33d51cd3 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.258168] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41f880a-c560-4572-987f-6dd1ffcf5544 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.266986] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5e6e7a-1695-4c95-8356-eb1a472eb6a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.298885] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a77fd9b-7b59-4084-a80b-d98b88942da0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.306679] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62f32a0-c213-406b-8424-6e7d1417216e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.320193] env[61768]: DEBUG nova.compute.provider_tree [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.343887] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.377178] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.531612] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.710116] env[61768]: DEBUG nova.compute.manager [req-1f35395f-33db-479c-963b-4fc9fc7c6f9c req-c8c65b4e-77e8-4240-aaac-92071701928b service nova] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Received event network-vif-deleted-90da7add-917d-486c-b346-1e3eb5b991a7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.823510] env[61768]: DEBUG nova.scheduler.client.report [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.032971] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Releasing lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.033424] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 609.033658] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 609.033985] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3e5872b-a6ba-4485-bace-7a1e06fcf2be {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.042943] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261feeda-d21b-42db-b6fb-c5b0ff6ca2b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.067991] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 21099fc3-dd66-4871-a87c-e0d2c56a7818 could not be found. [ 609.068454] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 609.069020] env[61768]: INFO nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Took 0.04 seconds to destroy the instance on the hypervisor. [ 609.069290] env[61768]: DEBUG oslo.service.loopingcall [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 609.069518] env[61768]: DEBUG nova.compute.manager [-] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.069615] env[61768]: DEBUG nova.network.neutron [-] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 609.095847] env[61768]: DEBUG nova.network.neutron [-] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.329074] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.004s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.329726] env[61768]: ERROR nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Traceback (most recent call last): [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self.driver.spawn(context, instance, image_meta, [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] vm_ref = self.build_virtual_machine(instance, [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.329726] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] for vif in network_info: [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] return self._sync_wrapper(fn, *args, **kwargs) [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self.wait() [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self[:] = self._gt.wait() [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] return self._exit_event.wait() [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] current.throw(*self._exc) [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.330080] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] result = function(*args, **kwargs) [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] return func(*args, **kwargs) [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] raise e [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] nwinfo = self.network_api.allocate_for_instance( [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] created_port_ids = self._update_ports_for_instance( [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] with excutils.save_and_reraise_exception(): [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] self.force_reraise() [ 609.330422] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] raise self.value [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] updated_port = self._update_port( [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] _ensure_no_port_binding_failure(port) [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] raise exception.PortBindingFailed(port_id=port['id']) [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] nova.exception.PortBindingFailed: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. [ 609.330714] env[61768]: ERROR nova.compute.manager [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] [ 609.330714] env[61768]: DEBUG nova.compute.utils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 609.331709] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.582s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 609.333356] env[61768]: INFO nova.compute.claims [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 609.336411] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Build of instance 0103eded-5a0b-4b7f-b98f-63d7866d501c was re-scheduled: Binding failed for port d337de5c-cbd4-4cea-8c96-f3f54122d90b, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 609.336880] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 609.337116] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Acquiring lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.337270] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Acquired lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.337432] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.357663] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.384463] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:52:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='82661f65-1cdc-4f17-b743-99fe34539fc7',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-475523417',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.384626] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.384777] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.384955] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.385120] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.385272] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.385497] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.386073] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.386265] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.386445] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.386652] env[61768]: DEBUG nova.virt.hardware [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.387763] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3a4f613-ed03-44b2-a289-69e6d089ffcb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.395592] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6cad65e-98c4-4c29-b8b6-92c561f963c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.446280] env[61768]: ERROR nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. [ 609.446280] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.446280] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.446280] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.446280] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.446280] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.446280] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.446280] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.446280] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.446280] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 609.446280] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.446280] env[61768]: ERROR nova.compute.manager raise self.value [ 609.446280] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.446280] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.446280] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.446280] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.446762] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.446762] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.446762] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. [ 609.446762] env[61768]: ERROR nova.compute.manager [ 609.446762] env[61768]: Traceback (most recent call last): [ 609.446762] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.446762] env[61768]: listener.cb(fileno) [ 609.446762] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.446762] env[61768]: result = function(*args, **kwargs) [ 609.446762] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.446762] env[61768]: return func(*args, **kwargs) [ 609.446762] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.446762] env[61768]: raise e [ 609.446762] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.446762] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 609.446762] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.446762] env[61768]: created_port_ids = self._update_ports_for_instance( [ 609.446762] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.446762] env[61768]: with excutils.save_and_reraise_exception(): [ 609.446762] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.446762] env[61768]: self.force_reraise() [ 609.446762] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.446762] env[61768]: raise self.value [ 609.446762] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.446762] env[61768]: updated_port = self._update_port( [ 609.446762] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.446762] env[61768]: _ensure_no_port_binding_failure(port) [ 609.446762] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.446762] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.447611] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. [ 609.447611] env[61768]: Removing descriptor: 14 [ 609.447611] env[61768]: ERROR nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Traceback (most recent call last): [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] yield resources [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self.driver.spawn(context, instance, image_meta, [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.447611] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] vm_ref = self.build_virtual_machine(instance, [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] for vif in network_info: [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return self._sync_wrapper(fn, *args, **kwargs) [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self.wait() [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self[:] = self._gt.wait() [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return self._exit_event.wait() [ 609.448071] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] result = hub.switch() [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return self.greenlet.switch() [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] result = function(*args, **kwargs) [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return func(*args, **kwargs) [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] raise e [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] nwinfo = self.network_api.allocate_for_instance( [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.448563] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] created_port_ids = self._update_ports_for_instance( [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] with excutils.save_and_reraise_exception(): [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self.force_reraise() [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] raise self.value [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] updated_port = self._update_port( [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] _ensure_no_port_binding_failure(port) [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.448928] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] raise exception.PortBindingFailed(port_id=port['id']) [ 609.449258] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] nova.exception.PortBindingFailed: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. [ 609.449258] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] [ 609.449258] env[61768]: INFO nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Terminating instance [ 609.450203] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.450336] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquired lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.450506] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 609.600739] env[61768]: DEBUG nova.network.neutron [-] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.857254] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.952972] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.968961] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.025811] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.103851] env[61768]: INFO nova.compute.manager [-] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Took 1.03 seconds to deallocate network for instance. [ 610.110153] env[61768]: DEBUG nova.compute.claims [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.110344] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.457790] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Releasing lock "refresh_cache-0103eded-5a0b-4b7f-b98f-63d7866d501c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.458172] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 610.458395] env[61768]: DEBUG nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.458602] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.476122] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.528572] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Releasing lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.528993] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 610.529208] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 610.531871] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed32f29e-8f90-4c63-bbe1-853d3719a499 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.540430] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ae072c-6669-45ee-b602-a7b78388fb65 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.565484] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 83afb27a-7984-4982-9880-f5b0c6553787 could not be found. [ 610.565706] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 610.565888] env[61768]: INFO nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Took 0.04 seconds to destroy the instance on the hypervisor. [ 610.566145] env[61768]: DEBUG oslo.service.loopingcall [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 610.568435] env[61768]: DEBUG nova.compute.manager [-] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 610.568582] env[61768]: DEBUG nova.network.neutron [-] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 610.582445] env[61768]: DEBUG nova.network.neutron [-] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 610.748696] env[61768]: DEBUG nova.compute.manager [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Received event network-changed-617cd2e1-7d6a-471a-a06a-973c33d51cd3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 610.748696] env[61768]: DEBUG nova.compute.manager [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Refreshing instance network info cache due to event network-changed-617cd2e1-7d6a-471a-a06a-973c33d51cd3. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 610.748696] env[61768]: DEBUG oslo_concurrency.lockutils [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] Acquiring lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.748696] env[61768]: DEBUG oslo_concurrency.lockutils [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] Acquired lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.748696] env[61768]: DEBUG nova.network.neutron [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Refreshing network info cache for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 610.752026] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab28ff5e-6379-4265-a208-359d6e67543c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.761025] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d22aac2-306a-4bc3-9687-fac6703b5321 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.790480] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ab2d33-44f2-464f-a4b5-f5db1e015cf1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.797823] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56e24ea9-f580-4af8-ad57-200429583780 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.811721] env[61768]: DEBUG nova.compute.provider_tree [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.980632] env[61768]: DEBUG nova.network.neutron [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.085676] env[61768]: DEBUG nova.network.neutron [-] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.264517] env[61768]: DEBUG nova.network.neutron [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.314753] env[61768]: DEBUG nova.scheduler.client.report [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.321326] env[61768]: DEBUG nova.network.neutron [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.483306] env[61768]: INFO nova.compute.manager [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] [instance: 0103eded-5a0b-4b7f-b98f-63d7866d501c] Took 1.02 seconds to deallocate network for instance. [ 611.588264] env[61768]: INFO nova.compute.manager [-] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Took 1.02 seconds to deallocate network for instance. [ 611.590928] env[61768]: DEBUG nova.compute.claims [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 611.591146] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 611.819478] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.819995] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.822630] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.247s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.825595] env[61768]: DEBUG oslo_concurrency.lockutils [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] Releasing lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.825762] env[61768]: DEBUG nova.compute.manager [req-957c494a-461d-4bdb-a5f4-b169422f983a req-e94758ec-c53f-4801-a150-e44b3e1d7ad7 service nova] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Received event network-vif-deleted-617cd2e1-7d6a-471a-a06a-973c33d51cd3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 612.330270] env[61768]: DEBUG nova.compute.utils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 612.331786] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 612.331986] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 612.369238] env[61768]: DEBUG nova.policy [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd26bd2d2c22848778d81d43853e1e5dc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '96e6619a79e54ce29d5e796ebac2338c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 612.516134] env[61768]: INFO nova.scheduler.client.report [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Deleted allocations for instance 0103eded-5a0b-4b7f-b98f-63d7866d501c [ 612.770782] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-027b8b63-761e-4a4e-88ed-3269afe00c7e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.774922] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Successfully created port: b17e938d-9211-4028-80a8-78251ed42bfa {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.782110] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7ec470e-54e8-4ef9-bd38-58c01b05df54 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.814802] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b43320-95d8-4ee9-9b03-2bdd8213ad37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.823075] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bf7ca1-5a7f-4fe2-9bb2-60ee242276f8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.836406] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.839468] env[61768]: DEBUG nova.compute.provider_tree [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.029165] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0ffebd87-363d-40e1-bd3b-67702bf89ed1 tempest-ServerRescueTestJSON-111826525 tempest-ServerRescueTestJSON-111826525-project-member] Lock "0103eded-5a0b-4b7f-b98f-63d7866d501c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.649s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.347248] env[61768]: DEBUG nova.scheduler.client.report [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.534121] env[61768]: DEBUG nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 613.851703] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.029s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.852543] env[61768]: ERROR nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Traceback (most recent call last): [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self.driver.spawn(context, instance, image_meta, [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] vm_ref = self.build_virtual_machine(instance, [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.852543] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] for vif in network_info: [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return self._sync_wrapper(fn, *args, **kwargs) [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self.wait() [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self[:] = self._gt.wait() [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return self._exit_event.wait() [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] result = hub.switch() [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.853042] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return self.greenlet.switch() [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] result = function(*args, **kwargs) [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] return func(*args, **kwargs) [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] raise e [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] nwinfo = self.network_api.allocate_for_instance( [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] created_port_ids = self._update_ports_for_instance( [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] with excutils.save_and_reraise_exception(): [ 613.853531] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] self.force_reraise() [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] raise self.value [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] updated_port = self._update_port( [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] _ensure_no_port_binding_failure(port) [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] raise exception.PortBindingFailed(port_id=port['id']) [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] nova.exception.PortBindingFailed: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. [ 613.854486] env[61768]: ERROR nova.compute.manager [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] [ 613.854891] env[61768]: DEBUG nova.compute.utils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 613.856083] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.858565] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Build of instance 70e6d944-f978-4d5b-b396-e0e913db0485 was re-scheduled: Binding failed for port 6cebe1cf-0a80-4fd5-8dac-209ebea16918, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 613.859017] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 613.859417] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquiring lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.859595] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Acquired lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.859760] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 613.861613] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.127s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.867120] env[61768]: DEBUG nova.compute.manager [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Received event network-changed-b17e938d-9211-4028-80a8-78251ed42bfa {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.867280] env[61768]: DEBUG nova.compute.manager [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Refreshing instance network info cache due to event network-changed-b17e938d-9211-4028-80a8-78251ed42bfa. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 613.867423] env[61768]: DEBUG oslo_concurrency.lockutils [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] Acquiring lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.867575] env[61768]: DEBUG oslo_concurrency.lockutils [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] Acquired lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.867735] env[61768]: DEBUG nova.network.neutron [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Refreshing network info cache for port b17e938d-9211-4028-80a8-78251ed42bfa {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 613.895023] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.895023] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.895023] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.895023] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.895268] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.895268] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.895268] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.895268] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.895762] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.896101] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.896387] env[61768]: DEBUG nova.virt.hardware [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.897360] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e986496-c93a-466d-8d4c-9aa1e72e1d70 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.908174] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02141b27-ccb3-4ab8-9a5b-24da8a4f422d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.057934] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.172784] env[61768]: ERROR nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. [ 614.172784] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.172784] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.172784] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.172784] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.172784] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.172784] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.172784] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.172784] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.172784] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 614.172784] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.172784] env[61768]: ERROR nova.compute.manager raise self.value [ 614.172784] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.172784] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.172784] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.172784] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.173511] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.173511] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.173511] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. [ 614.173511] env[61768]: ERROR nova.compute.manager [ 614.173511] env[61768]: Traceback (most recent call last): [ 614.173511] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.173511] env[61768]: listener.cb(fileno) [ 614.173511] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.173511] env[61768]: result = function(*args, **kwargs) [ 614.173511] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.173511] env[61768]: return func(*args, **kwargs) [ 614.173511] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.173511] env[61768]: raise e [ 614.173511] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.173511] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 614.173511] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.173511] env[61768]: created_port_ids = self._update_ports_for_instance( [ 614.173511] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.173511] env[61768]: with excutils.save_and_reraise_exception(): [ 614.173511] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.173511] env[61768]: self.force_reraise() [ 614.173511] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.173511] env[61768]: raise self.value [ 614.173511] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.173511] env[61768]: updated_port = self._update_port( [ 614.173511] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.173511] env[61768]: _ensure_no_port_binding_failure(port) [ 614.173511] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.173511] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.177268] env[61768]: nova.exception.PortBindingFailed: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. [ 614.177268] env[61768]: Removing descriptor: 14 [ 614.177268] env[61768]: ERROR nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Traceback (most recent call last): [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] yield resources [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self.driver.spawn(context, instance, image_meta, [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.177268] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] vm_ref = self.build_virtual_machine(instance, [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] for vif in network_info: [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return self._sync_wrapper(fn, *args, **kwargs) [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self.wait() [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self[:] = self._gt.wait() [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return self._exit_event.wait() [ 614.177728] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] result = hub.switch() [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return self.greenlet.switch() [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] result = function(*args, **kwargs) [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return func(*args, **kwargs) [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] raise e [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] nwinfo = self.network_api.allocate_for_instance( [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 614.178194] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] created_port_ids = self._update_ports_for_instance( [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] with excutils.save_and_reraise_exception(): [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self.force_reraise() [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] raise self.value [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] updated_port = self._update_port( [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] _ensure_no_port_binding_failure(port) [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.178539] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] raise exception.PortBindingFailed(port_id=port['id']) [ 614.178845] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] nova.exception.PortBindingFailed: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. [ 614.178845] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] [ 614.178845] env[61768]: INFO nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Terminating instance [ 614.178845] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Acquiring lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.396510] env[61768]: DEBUG nova.network.neutron [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.398712] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.535489] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.572868] env[61768]: DEBUG nova.network.neutron [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.793549] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17c651a-8ef9-4694-bbed-11847b2100a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.801619] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1a5681-5078-44e4-86aa-a3fd9c649528 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.832781] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d6774f6-9500-46aa-a5be-3a805f8f113f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.840303] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a23ee4a-2111-4079-a805-c48a1a019c49 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.855265] env[61768]: DEBUG nova.compute.provider_tree [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.038274] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Releasing lock "refresh_cache-70e6d944-f978-4d5b-b396-e0e913db0485" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.038525] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 615.038698] env[61768]: DEBUG nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.038868] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.054845] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.075543] env[61768]: DEBUG oslo_concurrency.lockutils [req-52dda9c4-3848-45d6-b535-95612d174eaf req-3d0c3af2-0792-4363-a82d-84c05633f374 service nova] Releasing lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.076069] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Acquired lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.076286] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.360102] env[61768]: DEBUG nova.scheduler.client.report [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.557615] env[61768]: DEBUG nova.network.neutron [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.612929] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.742444] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.866970] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.005s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.867703] env[61768]: ERROR nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Traceback (most recent call last): [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self.driver.spawn(context, instance, image_meta, [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] vm_ref = self.build_virtual_machine(instance, [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.867703] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] for vif in network_info: [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return self._sync_wrapper(fn, *args, **kwargs) [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self.wait() [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self[:] = self._gt.wait() [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return self._exit_event.wait() [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] result = hub.switch() [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.868109] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return self.greenlet.switch() [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] result = function(*args, **kwargs) [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] return func(*args, **kwargs) [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] raise e [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] nwinfo = self.network_api.allocate_for_instance( [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] created_port_ids = self._update_ports_for_instance( [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] with excutils.save_and_reraise_exception(): [ 615.868735] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] self.force_reraise() [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] raise self.value [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] updated_port = self._update_port( [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] _ensure_no_port_binding_failure(port) [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] raise exception.PortBindingFailed(port_id=port['id']) [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] nova.exception.PortBindingFailed: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. [ 615.869377] env[61768]: ERROR nova.compute.manager [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] [ 615.870115] env[61768]: DEBUG nova.compute.utils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.871426] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Build of instance fcddfd64-74bf-4a05-9799-0b7fee7416b2 was re-scheduled: Binding failed for port b43c8943-f264-45d4-8fcc-80670ee662e8, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 615.871943] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 615.872258] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.872452] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquired lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.872678] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.874469] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.469s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.874651] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.874803] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 615.875105] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.555s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.878827] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d97fed-d64b-432e-9a87-c3fced076ae7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.888250] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c50ae26e-a861-463a-8b0e-1b15863deb58 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.903677] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6377ae73-3caa-4722-a4ee-cc58f218fe6b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.912550] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f2c102-74c9-44f9-beb8-8a97045054f3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.918555] env[61768]: DEBUG nova.compute.manager [req-a964db41-ef90-4dd5-8493-ec325cbe6a72 req-06b17ef1-01c1-42a5-bde7-cb42c0df7e1c service nova] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Received event network-vif-deleted-b17e938d-9211-4028-80a8-78251ed42bfa {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.949271] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181450MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 615.949271] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.061463] env[61768]: INFO nova.compute.manager [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] [instance: 70e6d944-f978-4d5b-b396-e0e913db0485] Took 1.02 seconds to deallocate network for instance. [ 616.248022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Releasing lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.248022] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.248022] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 616.248022] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1713a4eb-3ed4-4ccd-a510-f088a0cba9d1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.258892] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98545ee-b906-4fdf-b9f6-84c1893f6283 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.292302] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8ba075db-5095-4c38-afbc-85e9d35d6194 could not be found. [ 616.292910] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 616.293337] env[61768]: INFO nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Took 0.05 seconds to destroy the instance on the hypervisor. [ 616.293879] env[61768]: DEBUG oslo.service.loopingcall [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.294317] env[61768]: DEBUG nova.compute.manager [-] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.294569] env[61768]: DEBUG nova.network.neutron [-] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.312589] env[61768]: DEBUG nova.network.neutron [-] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.407097] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.476804] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.819080] env[61768]: DEBUG nova.network.neutron [-] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.930172] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42f8c09a-9667-4b48-9d26-fdf8b1501638 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.945016] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4e1484-204c-484b-a14e-fad3951f0c43 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.983009] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Releasing lock "refresh_cache-fcddfd64-74bf-4a05-9799-0b7fee7416b2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.983301] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 616.983472] env[61768]: DEBUG nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.983672] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.986660] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5922d3-8081-4b05-a93f-0dacbedccce2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.995970] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-708b099a-7263-48b1-9298-46c5d3327203 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.010508] env[61768]: DEBUG nova.compute.provider_tree [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 617.013374] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 617.091552] env[61768]: INFO nova.scheduler.client.report [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Deleted allocations for instance 70e6d944-f978-4d5b-b396-e0e913db0485 [ 617.322717] env[61768]: INFO nova.compute.manager [-] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Took 1.03 seconds to deallocate network for instance. [ 617.326142] env[61768]: DEBUG nova.compute.claims [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.326142] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.520022] env[61768]: DEBUG nova.scheduler.client.report [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.520752] env[61768]: DEBUG nova.network.neutron [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.600099] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a8b4f1-3c2c-4705-a494-2bf3f4072965 tempest-DeleteServersAdminTestJSON-114849093 tempest-DeleteServersAdminTestJSON-114849093-project-member] Lock "70e6d944-f978-4d5b-b396-e0e913db0485" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.871s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.024070] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.149s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.024648] env[61768]: ERROR nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Traceback (most recent call last): [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self.driver.spawn(context, instance, image_meta, [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] vm_ref = self.build_virtual_machine(instance, [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] vif_infos = vmwarevif.get_vif_info(self._session, [ 618.024648] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] for vif in network_info: [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return self._sync_wrapper(fn, *args, **kwargs) [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self.wait() [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self[:] = self._gt.wait() [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return self._exit_event.wait() [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] result = hub.switch() [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 618.025377] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return self.greenlet.switch() [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] result = function(*args, **kwargs) [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] return func(*args, **kwargs) [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] raise e [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] nwinfo = self.network_api.allocate_for_instance( [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] created_port_ids = self._update_ports_for_instance( [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] with excutils.save_and_reraise_exception(): [ 618.025727] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] self.force_reraise() [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] raise self.value [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] updated_port = self._update_port( [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] _ensure_no_port_binding_failure(port) [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] raise exception.PortBindingFailed(port_id=port['id']) [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] nova.exception.PortBindingFailed: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. [ 618.026128] env[61768]: ERROR nova.compute.manager [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] [ 618.026433] env[61768]: DEBUG nova.compute.utils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 618.027030] env[61768]: INFO nova.compute.manager [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: fcddfd64-74bf-4a05-9799-0b7fee7416b2] Took 1.04 seconds to deallocate network for instance. [ 618.029727] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Build of instance f0be276a-ddad-4ae0-a232-d0fb8463a3bc was re-scheduled: Binding failed for port 8dda1026-0a57-4b2a-bbf4-7bd24c46175d, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 618.030231] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 618.030473] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Acquiring lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.030623] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Acquired lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.030779] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 618.031686] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.523s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.033281] env[61768]: INFO nova.compute.claims [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 618.106409] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 618.558785] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.667986] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.721253] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.947899] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.948188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.072765] env[61768]: INFO nova.scheduler.client.report [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Deleted allocations for instance fcddfd64-74bf-4a05-9799-0b7fee7416b2 [ 619.172724] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Releasing lock "refresh_cache-f0be276a-ddad-4ae0-a232-d0fb8463a3bc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.172966] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 619.173172] env[61768]: DEBUG nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.173338] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 619.191558] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 619.441502] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba9fa10f-d8b1-4afb-8a32-39b850a45d03 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.449668] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5733993e-9a30-41a5-9da4-7ae0da3953b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.485019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1d7287-214f-4dbb-8ae2-105d7b067761 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.490340] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab1e605-6eb0-4f8f-b129-301f2e0cc925 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.503723] env[61768]: DEBUG nova.compute.provider_tree [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 619.583643] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4514765-860e-41f8-9995-01edf9026ac3 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "fcddfd64-74bf-4a05-9799-0b7fee7416b2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.981s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.698580] env[61768]: DEBUG nova.network.neutron [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.010214] env[61768]: DEBUG nova.scheduler.client.report [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.086427] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 620.202978] env[61768]: INFO nova.compute.manager [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] [instance: f0be276a-ddad-4ae0-a232-d0fb8463a3bc] Took 1.03 seconds to deallocate network for instance. [ 620.516078] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 620.516515] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 620.519261] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.183s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.616403] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.024030] env[61768]: DEBUG nova.compute.utils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 621.028367] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 621.028700] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 621.078655] env[61768]: DEBUG nova.policy [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8b89f74739d8472c8727e9653c32d6a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '573bf2b9cd514369a3f9b429dd9c388e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 621.243528] env[61768]: INFO nova.scheduler.client.report [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Deleted allocations for instance f0be276a-ddad-4ae0-a232-d0fb8463a3bc [ 621.414526] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Successfully created port: 5bd4362c-14c9-4465-bef1-564b38320b12 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.446156] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b25698-84f2-4d92-93e6-28c844c33059 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.454516] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbfef87-c7ee-4531-9027-21a34a413e04 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.487998] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2626c3f-ebcc-425b-b1c8-716691dbfdd1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.495261] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb0ca5c-e3b9-4dce-a564-177d73397aa0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.509344] env[61768]: DEBUG nova.compute.provider_tree [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.533018] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 621.757668] env[61768]: DEBUG oslo_concurrency.lockutils [None req-88c1db0d-69ec-4b5a-bbe8-61fc792fbecc tempest-AttachInterfacesUnderV243Test-1071771707 tempest-AttachInterfacesUnderV243Test-1071771707-project-member] Lock "f0be276a-ddad-4ae0-a232-d0fb8463a3bc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.575s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.014101] env[61768]: DEBUG nova.scheduler.client.report [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.264296] env[61768]: DEBUG nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.266068] env[61768]: DEBUG nova.compute.manager [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Received event network-changed-5bd4362c-14c9-4465-bef1-564b38320b12 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 622.266505] env[61768]: DEBUG nova.compute.manager [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Refreshing instance network info cache due to event network-changed-5bd4362c-14c9-4465-bef1-564b38320b12. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 622.266908] env[61768]: DEBUG oslo_concurrency.lockutils [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] Acquiring lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.267277] env[61768]: DEBUG oslo_concurrency.lockutils [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] Acquired lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.267536] env[61768]: DEBUG nova.network.neutron [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Refreshing network info cache for port 5bd4362c-14c9-4465-bef1-564b38320b12 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 622.450619] env[61768]: ERROR nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. [ 622.450619] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 622.450619] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.450619] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 622.450619] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.450619] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 622.450619] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.450619] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 622.450619] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.450619] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 622.450619] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.450619] env[61768]: ERROR nova.compute.manager raise self.value [ 622.450619] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.450619] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 622.450619] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.450619] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 622.451062] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.451062] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 622.451062] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. [ 622.451062] env[61768]: ERROR nova.compute.manager [ 622.451062] env[61768]: Traceback (most recent call last): [ 622.451062] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 622.451062] env[61768]: listener.cb(fileno) [ 622.451062] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.451062] env[61768]: result = function(*args, **kwargs) [ 622.451062] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.451062] env[61768]: return func(*args, **kwargs) [ 622.451062] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.451062] env[61768]: raise e [ 622.451062] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.451062] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 622.451062] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.451062] env[61768]: created_port_ids = self._update_ports_for_instance( [ 622.451062] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.451062] env[61768]: with excutils.save_and_reraise_exception(): [ 622.451062] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.451062] env[61768]: self.force_reraise() [ 622.451062] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.451062] env[61768]: raise self.value [ 622.451062] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.451062] env[61768]: updated_port = self._update_port( [ 622.451062] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.451062] env[61768]: _ensure_no_port_binding_failure(port) [ 622.451062] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.451062] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 622.451865] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. [ 622.451865] env[61768]: Removing descriptor: 17 [ 622.519617] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.520441] env[61768]: ERROR nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Traceback (most recent call last): [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self.driver.spawn(context, instance, image_meta, [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] vm_ref = self.build_virtual_machine(instance, [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.520441] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] for vif in network_info: [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] return self._sync_wrapper(fn, *args, **kwargs) [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self.wait() [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self[:] = self._gt.wait() [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] return self._exit_event.wait() [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] current.throw(*self._exc) [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.520836] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] result = function(*args, **kwargs) [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] return func(*args, **kwargs) [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] raise e [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] nwinfo = self.network_api.allocate_for_instance( [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] created_port_ids = self._update_ports_for_instance( [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] with excutils.save_and_reraise_exception(): [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] self.force_reraise() [ 622.521161] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] raise self.value [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] updated_port = self._update_port( [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] _ensure_no_port_binding_failure(port) [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] raise exception.PortBindingFailed(port_id=port['id']) [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] nova.exception.PortBindingFailed: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. [ 622.521464] env[61768]: ERROR nova.compute.manager [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] [ 622.522530] env[61768]: DEBUG nova.compute.utils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.524092] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.488s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.525766] env[61768]: INFO nova.compute.claims [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 622.531021] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Build of instance aadc35e7-7aea-4f86-be3a-9b5322f350a5 was re-scheduled: Binding failed for port de30f4a1-b25d-44ab-a19e-8685a941cc60, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.531021] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.531021] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquiring lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.531021] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Acquired lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.531318] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.544019] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 622.582116] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 622.582116] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 622.582116] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 622.582295] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 622.582295] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 622.582864] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 622.583246] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 622.583567] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 622.583869] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 622.584167] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 622.584667] env[61768]: DEBUG nova.virt.hardware [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.586606] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1877669-05d4-4d90-8889-638a744d2c80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.596569] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c3033e9-029f-4e2f-882f-0b52207a9c5a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.612096] env[61768]: ERROR nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Traceback (most recent call last): [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] yield resources [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self.driver.spawn(context, instance, image_meta, [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] vm_ref = self.build_virtual_machine(instance, [ 622.612096] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] for vif in network_info: [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] return self._sync_wrapper(fn, *args, **kwargs) [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self.wait() [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self[:] = self._gt.wait() [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] return self._exit_event.wait() [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 622.612449] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] current.throw(*self._exc) [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] result = function(*args, **kwargs) [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] return func(*args, **kwargs) [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] raise e [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] nwinfo = self.network_api.allocate_for_instance( [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] created_port_ids = self._update_ports_for_instance( [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] with excutils.save_and_reraise_exception(): [ 622.612871] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self.force_reraise() [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] raise self.value [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] updated_port = self._update_port( [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] _ensure_no_port_binding_failure(port) [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] raise exception.PortBindingFailed(port_id=port['id']) [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] nova.exception.PortBindingFailed: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. [ 622.613237] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] [ 622.613570] env[61768]: INFO nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Terminating instance [ 622.619205] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Acquiring lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.793921] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.802114] env[61768]: DEBUG nova.network.neutron [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.969412] env[61768]: DEBUG nova.network.neutron [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.073425] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.163759] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.473625] env[61768]: DEBUG oslo_concurrency.lockutils [req-ba061c91-b888-4fb3-a34f-69b82a1bcaa8 req-ff9b3b30-47e0-41a1-91e5-223d893f63b3 service nova] Releasing lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.473903] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Acquired lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.474032] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 623.666663] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Releasing lock "refresh_cache-aadc35e7-7aea-4f86-be3a-9b5322f350a5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.666948] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.667133] env[61768]: DEBUG nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.667302] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.683962] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.974089] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662f5f4e-19d5-4473-8cd8-b29092e4e52c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.988490] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d3c0aa-c104-4b71-b81d-29239cc355e8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.031889] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.032718] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e6eca0-a03a-4c50-b92e-d20a2ddff2c6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.040684] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b36cf9b9-72c3-4e10-a078-7440022dc721 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.055930] env[61768]: DEBUG nova.compute.provider_tree [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.144886] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.185656] env[61768]: DEBUG nova.network.neutron [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.325058] env[61768]: DEBUG nova.compute.manager [req-2c722b90-069f-44b6-a1f4-3462a7075e7d req-4fd358b0-f2f3-49d6-8a35-9745b8c306bf service nova] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Received event network-vif-deleted-5bd4362c-14c9-4465-bef1-564b38320b12 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.563013] env[61768]: DEBUG nova.scheduler.client.report [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.647459] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Releasing lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.647880] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.648444] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 624.648444] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-996955a4-0515-4bf3-9018-3eb1e22eb063 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.657419] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd31661d-d7ad-46f1-b8bb-b884fefb62f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.679172] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e6922bb7-c2df-468a-aa5b-d03e274b8ecc could not be found. [ 624.679409] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 624.681057] env[61768]: INFO nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Took 0.03 seconds to destroy the instance on the hypervisor. [ 624.681057] env[61768]: DEBUG oslo.service.loopingcall [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.681057] env[61768]: DEBUG nova.compute.manager [-] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.681057] env[61768]: DEBUG nova.network.neutron [-] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 624.691660] env[61768]: INFO nova.compute.manager [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] [instance: aadc35e7-7aea-4f86-be3a-9b5322f350a5] Took 1.02 seconds to deallocate network for instance. [ 624.700043] env[61768]: DEBUG nova.network.neutron [-] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.074417] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.074949] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 625.078630] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.967s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.205537] env[61768]: DEBUG nova.network.neutron [-] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.583074] env[61768]: DEBUG nova.compute.utils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 625.585820] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 625.585961] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 625.646713] env[61768]: DEBUG nova.policy [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1bc181186137437bb45cd48aee6ee5c6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1ba4e32b9f6d42ebb2a0701ef74da19f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 625.712787] env[61768]: INFO nova.compute.manager [-] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Took 1.03 seconds to deallocate network for instance. [ 625.717873] env[61768]: DEBUG nova.compute.claims [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.718090] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 625.727610] env[61768]: INFO nova.scheduler.client.report [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Deleted allocations for instance aadc35e7-7aea-4f86-be3a-9b5322f350a5 [ 626.019822] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e9bec3-dd4c-4f83-b528-e5a551f8d68b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.028854] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb4d485c-771a-4377-8c09-fd9969625c67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.063320] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37349fdd-608c-4b07-9488-ee90117d5bf4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.071545] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f522ad29-90c0-497e-9e0b-538970ece1a8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.084799] env[61768]: DEBUG nova.compute.provider_tree [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.086220] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 626.101163] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Successfully created port: 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 626.236748] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fc7f82a8-7b2b-4aa8-bcbe-ed912aeba458 tempest-ListImageFiltersTestJSON-601317255 tempest-ListImageFiltersTestJSON-601317255-project-member] Lock "aadc35e7-7aea-4f86-be3a-9b5322f350a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 94.683s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.269257] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.269440] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.294022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "9fa32a4d-3ba4-4d36-963b-17a64453e804" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.294755] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.590949] env[61768]: DEBUG nova.scheduler.client.report [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.743450] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.070103] env[61768]: DEBUG nova.compute.manager [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Received event network-changed-6b219d2a-ed79-4bfd-b45b-5ea6d8d90073 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.070103] env[61768]: DEBUG nova.compute.manager [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Refreshing instance network info cache due to event network-changed-6b219d2a-ed79-4bfd-b45b-5ea6d8d90073. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 627.070103] env[61768]: DEBUG oslo_concurrency.lockutils [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] Acquiring lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.070103] env[61768]: DEBUG oslo_concurrency.lockutils [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] Acquired lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.070493] env[61768]: DEBUG nova.network.neutron [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Refreshing network info cache for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 627.102899] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.025s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.103923] env[61768]: ERROR nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Traceback (most recent call last): [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self.driver.spawn(context, instance, image_meta, [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] vm_ref = self.build_virtual_machine(instance, [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.103923] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] for vif in network_info: [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return self._sync_wrapper(fn, *args, **kwargs) [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self.wait() [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self[:] = self._gt.wait() [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return self._exit_event.wait() [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] result = hub.switch() [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 627.104401] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return self.greenlet.switch() [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] result = function(*args, **kwargs) [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] return func(*args, **kwargs) [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] raise e [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] nwinfo = self.network_api.allocate_for_instance( [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] created_port_ids = self._update_ports_for_instance( [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] with excutils.save_and_reraise_exception(): [ 627.104883] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] self.force_reraise() [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] raise self.value [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] updated_port = self._update_port( [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] _ensure_no_port_binding_failure(port) [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] raise exception.PortBindingFailed(port_id=port['id']) [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] nova.exception.PortBindingFailed: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. [ 627.105689] env[61768]: ERROR nova.compute.manager [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] [ 627.106222] env[61768]: DEBUG nova.compute.utils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 627.108876] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 627.114194] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Build of instance 21099fc3-dd66-4871-a87c-e0d2c56a7818 was re-scheduled: Binding failed for port 90da7add-917d-486c-b346-1e3eb5b991a7, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 627.114194] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 627.114194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Acquiring lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.114194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Acquired lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.114345] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 627.114971] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.524s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.152522] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 627.152746] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 627.152922] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 627.153162] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 627.153715] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 627.153715] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 627.153715] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 627.153859] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 627.153979] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 627.154643] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 627.155057] env[61768]: DEBUG nova.virt.hardware [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 627.156100] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31225946-8f0d-4e64-82e2-0c099fe1679b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.165087] env[61768]: ERROR nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. [ 627.165087] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 627.165087] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.165087] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 627.165087] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.165087] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 627.165087] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.165087] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 627.165087] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.165087] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 627.165087] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.165087] env[61768]: ERROR nova.compute.manager raise self.value [ 627.165087] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.165087] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 627.165087] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.165087] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 627.165631] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.165631] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 627.165631] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. [ 627.165631] env[61768]: ERROR nova.compute.manager [ 627.165631] env[61768]: Traceback (most recent call last): [ 627.165631] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 627.165631] env[61768]: listener.cb(fileno) [ 627.165631] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.165631] env[61768]: result = function(*args, **kwargs) [ 627.165631] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.165631] env[61768]: return func(*args, **kwargs) [ 627.165631] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.165631] env[61768]: raise e [ 627.165631] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.165631] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 627.165631] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.165631] env[61768]: created_port_ids = self._update_ports_for_instance( [ 627.165631] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.165631] env[61768]: with excutils.save_and_reraise_exception(): [ 627.165631] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.165631] env[61768]: self.force_reraise() [ 627.165631] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.165631] env[61768]: raise self.value [ 627.165631] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.165631] env[61768]: updated_port = self._update_port( [ 627.165631] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.165631] env[61768]: _ensure_no_port_binding_failure(port) [ 627.165631] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.165631] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 627.166379] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. [ 627.166379] env[61768]: Removing descriptor: 17 [ 627.166568] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b45b0ea-95c4-479e-8562-9f9a81676c50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.188328] env[61768]: ERROR nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Traceback (most recent call last): [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] yield resources [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self.driver.spawn(context, instance, image_meta, [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self._vmops.spawn(context, instance, image_meta, injected_files, [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] vm_ref = self.build_virtual_machine(instance, [ 627.188328] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] vif_infos = vmwarevif.get_vif_info(self._session, [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] for vif in network_info: [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] return self._sync_wrapper(fn, *args, **kwargs) [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self.wait() [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self[:] = self._gt.wait() [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] return self._exit_event.wait() [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 627.188642] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] current.throw(*self._exc) [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] result = function(*args, **kwargs) [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] return func(*args, **kwargs) [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] raise e [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] nwinfo = self.network_api.allocate_for_instance( [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] created_port_ids = self._update_ports_for_instance( [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] with excutils.save_and_reraise_exception(): [ 627.188975] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self.force_reraise() [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] raise self.value [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] updated_port = self._update_port( [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] _ensure_no_port_binding_failure(port) [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] raise exception.PortBindingFailed(port_id=port['id']) [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] nova.exception.PortBindingFailed: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. [ 627.189304] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] [ 627.192138] env[61768]: INFO nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Terminating instance [ 627.192830] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Acquiring lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 627.270365] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.592854] env[61768]: DEBUG nova.network.neutron [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.652125] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 627.789106] env[61768]: DEBUG nova.network.neutron [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.823546] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.101954] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c33d91-eacc-4a00-950d-7f5c56c2581d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.109775] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2071a807-fed4-4181-873d-90073a11b7bf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.139783] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6d4c61-6f5c-4f65-a8a5-583468d52b1c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.147213] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72a9ea5-5a7e-4f5c-8ba3-1b67bea729ba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.163869] env[61768]: DEBUG nova.compute.provider_tree [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.292786] env[61768]: DEBUG oslo_concurrency.lockutils [req-9863993f-a0c7-4c39-9398-d35ff6364c22 req-fc82285c-cf57-4ec0-8a6b-c8ce83c3981f service nova] Releasing lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.292786] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Acquired lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.292986] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 628.326067] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Releasing lock "refresh_cache-21099fc3-dd66-4871-a87c-e0d2c56a7818" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.326683] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 628.326683] env[61768]: DEBUG nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.326683] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 628.354542] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.666948] env[61768]: DEBUG nova.scheduler.client.report [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.814813] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 628.857374] env[61768]: DEBUG nova.network.neutron [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 628.887072] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "9b3632e5-4c2d-4968-9692-70f91883bfb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.887234] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.132170] env[61768]: DEBUG nova.compute.manager [req-43261842-aba1-4bf0-a9e1-334a52faad32 req-fdc2035f-9dd2-4964-b8ee-dc4dbe74d708 service nova] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Received event network-vif-deleted-6b219d2a-ed79-4bfd-b45b-5ea6d8d90073 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 629.140545] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.172841] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.058s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.173653] env[61768]: ERROR nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Traceback (most recent call last): [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self.driver.spawn(context, instance, image_meta, [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] vm_ref = self.build_virtual_machine(instance, [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.173653] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] for vif in network_info: [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return self._sync_wrapper(fn, *args, **kwargs) [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self.wait() [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self[:] = self._gt.wait() [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return self._exit_event.wait() [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] result = hub.switch() [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.173983] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return self.greenlet.switch() [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] result = function(*args, **kwargs) [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] return func(*args, **kwargs) [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] raise e [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] nwinfo = self.network_api.allocate_for_instance( [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] created_port_ids = self._update_ports_for_instance( [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] with excutils.save_and_reraise_exception(): [ 629.174303] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] self.force_reraise() [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] raise self.value [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] updated_port = self._update_port( [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] _ensure_no_port_binding_failure(port) [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] raise exception.PortBindingFailed(port_id=port['id']) [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] nova.exception.PortBindingFailed: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. [ 629.174668] env[61768]: ERROR nova.compute.manager [instance: 83afb27a-7984-4982-9880-f5b0c6553787] [ 629.174938] env[61768]: DEBUG nova.compute.utils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.176341] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Build of instance 83afb27a-7984-4982-9880-f5b0c6553787 was re-scheduled: Binding failed for port 617cd2e1-7d6a-471a-a06a-973c33d51cd3, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.176940] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.177213] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquiring lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.177364] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Acquired lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.177524] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 629.180069] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.121s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.180496] env[61768]: INFO nova.compute.claims [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.361474] env[61768]: INFO nova.compute.manager [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] [instance: 21099fc3-dd66-4871-a87c-e0d2c56a7818] Took 1.03 seconds to deallocate network for instance. [ 629.643695] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Releasing lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.644489] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.644748] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 629.645109] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0fd774e7-a9b3-41d1-807a-4bdb8d47c46d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.654527] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9170dd3-cd34-48f9-bafc-e7773f7cd78b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.681031] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 07d8d6b6-3cc5-48da-9d4d-563a5fe27442 could not be found. [ 629.681298] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 629.681490] env[61768]: INFO nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Took 0.04 seconds to destroy the instance on the hypervisor. [ 629.681745] env[61768]: DEBUG oslo.service.loopingcall [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.681962] env[61768]: DEBUG nova.compute.manager [-] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.682064] env[61768]: DEBUG nova.network.neutron [-] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 629.709616] env[61768]: DEBUG nova.network.neutron [-] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.721150] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 629.813398] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.211899] env[61768]: DEBUG nova.network.neutron [-] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.320670] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Releasing lock "refresh_cache-83afb27a-7984-4982-9880-f5b0c6553787" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.320670] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 630.320670] env[61768]: DEBUG nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 630.320670] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 630.342390] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.395324] env[61768]: INFO nova.scheduler.client.report [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Deleted allocations for instance 21099fc3-dd66-4871-a87c-e0d2c56a7818 [ 630.668110] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e86dc8-b0a0-4a6d-ad33-2e6099d68b92 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.677333] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-821d9ac8-8398-45da-a089-26b29bca39ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.709920] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b83fc709-3282-4522-bcb2-6ea2bdc68301 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.717813] env[61768]: INFO nova.compute.manager [-] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Took 1.04 seconds to deallocate network for instance. [ 630.720745] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c08935-4e16-4806-b632-a3d3b6588656 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.725120] env[61768]: DEBUG nova.compute.claims [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.725252] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.735381] env[61768]: DEBUG nova.compute.provider_tree [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.849930] env[61768]: DEBUG nova.network.neutron [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.906632] env[61768]: DEBUG oslo_concurrency.lockutils [None req-36cb8640-1d08-466e-9198-5ec0413ab1ba tempest-ImagesOneServerTestJSON-1093653889 tempest-ImagesOneServerTestJSON-1093653889-project-member] Lock "21099fc3-dd66-4871-a87c-e0d2c56a7818" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.322s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.238093] env[61768]: DEBUG nova.scheduler.client.report [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.352528] env[61768]: INFO nova.compute.manager [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] [instance: 83afb27a-7984-4982-9880-f5b0c6553787] Took 1.03 seconds to deallocate network for instance. [ 631.409858] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 631.745584] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.567s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.746185] env[61768]: DEBUG nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 631.748890] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.800s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.935054] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.253480] env[61768]: DEBUG nova.compute.utils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 632.257650] env[61768]: DEBUG nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Not allocating networking since 'none' was specified. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 632.386019] env[61768]: INFO nova.scheduler.client.report [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Deleted allocations for instance 83afb27a-7984-4982-9880-f5b0c6553787 [ 632.758348] env[61768]: DEBUG nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.792459] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 8ba075db-5095-4c38-afbc-85e9d35d6194 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 632.792630] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance e6922bb7-c2df-468a-aa5b-d03e274b8ecc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 632.792758] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 07d8d6b6-3cc5-48da-9d4d-563a5fe27442 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 632.792882] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance f81aabf6-ae46-405c-b101-12ca707a0567 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 632.903476] env[61768]: DEBUG oslo_concurrency.lockutils [None req-02f87039-411b-4bf1-86af-153dae821b42 tempest-MigrationsAdminTest-1933659061 tempest-MigrationsAdminTest-1933659061-project-member] Lock "83afb27a-7984-4982-9880-f5b0c6553787" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.504s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.296426] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.404943] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 633.773233] env[61768]: DEBUG nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 633.802516] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 633.802767] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 633.802928] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 633.803122] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 633.803276] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 633.803427] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 633.803636] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 633.803796] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 633.803967] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 633.804154] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 633.804330] env[61768]: DEBUG nova.virt.hardware [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.805026] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7d35cf35-d264-468d-a9ab-7b41dd020023 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 633.806911] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ebaba7-ad37-4c6a-848d-a993630741d8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.816474] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdab3014-eba2-4ef0-a574-9401bfefd5a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.831293] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 633.836802] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Creating folder: Project (d92bc16050144333b255cb94fd15fe68). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 633.837368] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6d0d1aae-175f-4b30-877a-20caccf56bda {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.847457] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Created folder: Project (d92bc16050144333b255cb94fd15fe68) in parent group-v265360. [ 633.847457] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Creating folder: Instances. Parent ref: group-v265369. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 633.849027] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1d7154c6-c264-4d42-a16f-5faea0251e15 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.859498] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Created folder: Instances in parent group-v265369. [ 633.859498] env[61768]: DEBUG oslo.service.loopingcall [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.859498] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 633.859498] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5aa42705-115f-4e42-b410-5e82c04559e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.881267] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 633.881267] env[61768]: value = "task-1228559" [ 633.881267] env[61768]: _type = "Task" [ 633.881267] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.889078] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228559, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.939628] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.313465] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 1eb76ef5-377c-4447-92da-4c61345c0070 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.394086] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228559, 'name': CreateVM_Task, 'duration_secs': 0.280354} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.394276] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 634.394945] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.395138] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.395631] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 634.395994] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9ef8351-e8ec-4f32-8dfc-1a37f59510f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.401819] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 634.401819] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bccddb-f5e4-c375-b681-1e0f8ec95a94" [ 634.401819] env[61768]: _type = "Task" [ 634.401819] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.409931] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bccddb-f5e4-c375-b681-1e0f8ec95a94, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.816802] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 6035033d-9e22-49bd-a42c-3faeeb7c5d21 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 634.916412] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bccddb-f5e4-c375-b681-1e0f8ec95a94, 'name': SearchDatastore_Task, 'duration_secs': 0.010269} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.916791] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.917105] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 634.917373] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 634.917524] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.917990] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 634.918302] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-731098f7-a19b-4af6-a503-0399a3ac6980 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.927804] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 634.928310] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 634.928953] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b810f975-bf13-4c98-b5be-6d3a76560a12 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.935687] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 634.935687] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52717d6b-f21b-6598-83a4-b0bc47e917f9" [ 634.935687] env[61768]: _type = "Task" [ 634.935687] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.944230] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52717d6b-f21b-6598-83a4-b0bc47e917f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.321505] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 1def011b-674f-4336-967c-96d26d48aa6c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.447715] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52717d6b-f21b-6598-83a4-b0bc47e917f9, 'name': SearchDatastore_Task, 'duration_secs': 0.01617} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.448550] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05bd01a9-1ff3-4c63-aca7-3715e0989610 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.455244] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 635.455244] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5289c03e-ad3e-2209-6ed1-f6bb64b97399" [ 635.455244] env[61768]: _type = "Task" [ 635.455244] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.466860] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5289c03e-ad3e-2209-6ed1-f6bb64b97399, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.826628] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 496c3799-7350-4258-a456-58bd3ba9b4d0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 635.966614] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5289c03e-ad3e-2209-6ed1-f6bb64b97399, 'name': SearchDatastore_Task, 'duration_secs': 0.015125} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.966939] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.967215] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] f81aabf6-ae46-405c-b101-12ca707a0567/f81aabf6-ae46-405c-b101-12ca707a0567.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 635.967481] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0417082e-5962-4859-bb1c-18e9be742792 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.974175] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 635.974175] env[61768]: value = "task-1228560" [ 635.974175] env[61768]: _type = "Task" [ 635.974175] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.983112] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228560, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.329764] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance aa0833ae-99c7-426d-a972-8183667a48f8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 636.489887] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228560, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483147} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 636.489887] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] f81aabf6-ae46-405c-b101-12ca707a0567/f81aabf6-ae46-405c-b101-12ca707a0567.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 636.489887] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 636.489887] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45d8c6f2-a505-4359-a3c7-d2ee201ab473 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.496983] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 636.496983] env[61768]: value = "task-1228561" [ 636.496983] env[61768]: _type = "Task" [ 636.496983] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 636.507900] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228561, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 636.833027] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 6540f636-fdfd-48aa-a7b8-0bbc231741fe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.009893] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228561, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062498} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 637.010202] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 637.010977] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82341ba5-cab2-4b67-a311-37dc3afb1a68 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.031057] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Reconfiguring VM instance instance-0000001e to attach disk [datastore1] f81aabf6-ae46-405c-b101-12ca707a0567/f81aabf6-ae46-405c-b101-12ca707a0567.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 637.031299] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaac8609-375e-4f3a-ae8a-5b0e6088c8e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.052164] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 637.052164] env[61768]: value = "task-1228562" [ 637.052164] env[61768]: _type = "Task" [ 637.052164] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.060504] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228562, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.336600] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 5da49801-148d-4b26-a0a2-816b1d5f7994 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 637.486608] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "23a4450d-162a-4b2b-a009-7023851315a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.486820] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "23a4450d-162a-4b2b-a009-7023851315a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.564864] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228562, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.842235] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 63704075-5eaf-4f84-a90b-3a0a3e904a9d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.061808] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228562, 'name': ReconfigVM_Task, 'duration_secs': 0.779148} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.062869] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Reconfigured VM instance instance-0000001e to attach disk [datastore1] f81aabf6-ae46-405c-b101-12ca707a0567/f81aabf6-ae46-405c-b101-12ca707a0567.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 638.062869] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a39030a-f648-4613-8026-e93594c1548f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.070803] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 638.070803] env[61768]: value = "task-1228563" [ 638.070803] env[61768]: _type = "Task" [ 638.070803] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.077884] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228563, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.345054] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance c299abc5-7913-4eea-ac0a-02e713c6009c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 638.403117] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "d70040f1-e425-4aeb-b82d-ce808a02a645" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 638.403368] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.579180] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228563, 'name': Rename_Task, 'duration_secs': 0.139819} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.579529] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 638.579849] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-08c16e6e-9fbd-477c-a61b-86d30d4ced15 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.586467] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 638.586467] env[61768]: value = "task-1228564" [ 638.586467] env[61768]: _type = "Task" [ 638.586467] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.593965] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228564, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.850390] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance a83f5b0f-508c-4829-a4a6-641baa7b4c95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.095768] env[61768]: DEBUG oslo_vmware.api [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228564, 'name': PowerOnVM_Task, 'duration_secs': 0.508014} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.096073] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 639.096281] env[61768]: INFO nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Took 5.32 seconds to spawn the instance on the hypervisor. [ 639.096520] env[61768]: DEBUG nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 639.097509] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2226a221-87cb-44a7-a193-e74fe34ab529 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.355283] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance f6b07607-632d-46d9-a72f-5a524fd20a3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 639.615085] env[61768]: INFO nova.compute.manager [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Took 25.57 seconds to build instance. [ 639.857854] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 12a543c2-1081-49d7-800b-07f0a2516904 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.117675] env[61768]: DEBUG oslo_concurrency.lockutils [None req-931bb042-7be0-43a1-bce8-c07ea287c146 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "f81aabf6-ae46-405c-b101-12ca707a0567" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 100.776s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.360978] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 640.620876] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.869095] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.144611] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.371476] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 0df161e9-4a66-432e-9366-60f231c53e3d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 641.874669] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 3401be45-c3ab-4780-81a7-a93b05742414 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.379842] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 91d56a8c-0664-4b3c-bf28-e2c668ac65fa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 642.883210] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance efbde802-8cb0-4563-a776-3722a3889afe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.387029] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 643.891185] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 668d4fe4-85f8-4282-8d65-0549d9e3bda8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.396901] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 9fa32a4d-3ba4-4d36-963b-17a64453e804 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.898710] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 9b3632e5-4c2d-4968-9692-70f91883bfb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 644.899836] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 644.899836] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 645.328664] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b569a6-2930-4308-8702-211a847fc066 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.336335] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd3c64e6-74d4-427f-adca-c13403baa95f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.373021] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81482214-b098-4d7d-a7e1-d46323b91d2c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.379173] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dcb376f-8cc2-4f0d-ab25-aab19f3ab3b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.392344] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.895031] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.400887] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 646.401256] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.652s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.401550] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.075s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.676836] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.677111] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.168078] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.168320] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.343622] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dd61a74-d5dd-4cb0-b57e-1d98eb8da6fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.351469] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c53dd8-8a39-4958-9206-70f61ef7589b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.381327] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e214ae98-a560-464c-92f3-a6b4f48508e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.389431] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-830d73da-a376-4cf4-87f0-dbfe770581b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.402903] env[61768]: DEBUG nova.compute.provider_tree [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.908793] env[61768]: DEBUG nova.scheduler.client.report [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.417731] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.016s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.418457] env[61768]: ERROR nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Traceback (most recent call last): [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self.driver.spawn(context, instance, image_meta, [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] vm_ref = self.build_virtual_machine(instance, [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.418457] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] for vif in network_info: [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return self._sync_wrapper(fn, *args, **kwargs) [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self.wait() [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self[:] = self._gt.wait() [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return self._exit_event.wait() [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] result = hub.switch() [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.418824] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return self.greenlet.switch() [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] result = function(*args, **kwargs) [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] return func(*args, **kwargs) [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] raise e [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] nwinfo = self.network_api.allocate_for_instance( [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] created_port_ids = self._update_ports_for_instance( [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] with excutils.save_and_reraise_exception(): [ 648.419204] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] self.force_reraise() [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] raise self.value [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] updated_port = self._update_port( [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] _ensure_no_port_binding_failure(port) [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] raise exception.PortBindingFailed(port_id=port['id']) [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] nova.exception.PortBindingFailed: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. [ 648.419569] env[61768]: ERROR nova.compute.manager [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] [ 648.419886] env[61768]: DEBUG nova.compute.utils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.420480] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.699s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.421935] env[61768]: INFO nova.compute.claims [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.426149] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Build of instance 8ba075db-5095-4c38-afbc-85e9d35d6194 was re-scheduled: Binding failed for port b17e938d-9211-4028-80a8-78251ed42bfa, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 648.426756] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 648.427018] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Acquiring lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.427175] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Acquired lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.427339] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.944053] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.014493] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.517969] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Releasing lock "refresh_cache-8ba075db-5095-4c38-afbc-85e9d35d6194" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.518285] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 649.518429] env[61768]: DEBUG nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.518570] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 649.533378] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.799523] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c59a685-704e-4212-afb3-59a638bcf563 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.807836] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d5493d-0296-4c52-8097-0cfcb853dc14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.841628] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1334c5a2-214c-4aa3-bd16-44cc112749a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.849594] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b311fb2-52ba-4040-9cad-71fa44b65823 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.865198] env[61768]: DEBUG nova.compute.provider_tree [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.037020] env[61768]: DEBUG nova.network.neutron [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.371669] env[61768]: DEBUG nova.scheduler.client.report [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.538771] env[61768]: INFO nova.compute.manager [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] [instance: 8ba075db-5095-4c38-afbc-85e9d35d6194] Took 1.02 seconds to deallocate network for instance. [ 650.877213] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.877786] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.880599] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.264s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.881944] env[61768]: INFO nova.compute.claims [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 651.386503] env[61768]: DEBUG nova.compute.utils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 651.389841] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 651.390559] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 651.429172] env[61768]: DEBUG nova.policy [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '859e4f1eabe44fa3a9b9901c727da22f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0921db90fd746d094d31c2b0f6e5e72', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.568387] env[61768]: INFO nova.scheduler.client.report [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Deleted allocations for instance 8ba075db-5095-4c38-afbc-85e9d35d6194 [ 651.896718] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.916776] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Successfully created port: e9f4b2f1-f86d-4331-8c47-42dbc1299a0e {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.080310] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8b25ee18-9b0e-4ff4-9430-93fdfe20e80f tempest-ServersTestJSON-1979620588 tempest-ServersTestJSON-1979620588-project-member] Lock "8ba075db-5095-4c38-afbc-85e9d35d6194" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.916s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.292579] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fc3bd4-b78f-4f1c-a3d7-5d12bd00efa2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.300207] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54ff3742-e132-479d-83d8-ecac07653575 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.331427] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212aa755-9246-4922-9688-47abfd3c44e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.338377] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32617922-e4b6-4458-a860-aef12919b9dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.350855] env[61768]: DEBUG nova.compute.provider_tree [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.584167] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 652.812114] env[61768]: DEBUG nova.compute.manager [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Received event network-changed-e9f4b2f1-f86d-4331-8c47-42dbc1299a0e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.812114] env[61768]: DEBUG nova.compute.manager [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Refreshing instance network info cache due to event network-changed-e9f4b2f1-f86d-4331-8c47-42dbc1299a0e. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 652.812114] env[61768]: DEBUG oslo_concurrency.lockutils [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] Acquiring lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.812114] env[61768]: DEBUG oslo_concurrency.lockutils [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] Acquired lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.812114] env[61768]: DEBUG nova.network.neutron [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Refreshing network info cache for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 652.854035] env[61768]: DEBUG nova.scheduler.client.report [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.906700] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.936519] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.936775] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.937016] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.937253] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.937406] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.937557] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.938014] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.938280] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.938520] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.938736] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.939066] env[61768]: DEBUG nova.virt.hardware [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.940196] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6bc59b1-147d-438e-89c1-c13b6ed5ab40 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.945437] env[61768]: ERROR nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. [ 652.945437] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.945437] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.945437] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.945437] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.945437] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.945437] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.945437] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.945437] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.945437] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 652.945437] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.945437] env[61768]: ERROR nova.compute.manager raise self.value [ 652.945437] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.945437] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.945437] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.945437] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.945929] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.945929] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.945929] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. [ 652.945929] env[61768]: ERROR nova.compute.manager [ 652.945929] env[61768]: Traceback (most recent call last): [ 652.946114] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.946114] env[61768]: listener.cb(fileno) [ 652.946114] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.946114] env[61768]: result = function(*args, **kwargs) [ 652.946114] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.946114] env[61768]: return func(*args, **kwargs) [ 652.946114] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.946114] env[61768]: raise e [ 652.946114] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.946114] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 652.946114] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.946114] env[61768]: created_port_ids = self._update_ports_for_instance( [ 652.946114] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.946114] env[61768]: with excutils.save_and_reraise_exception(): [ 652.946114] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.946114] env[61768]: self.force_reraise() [ 652.946114] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.946114] env[61768]: raise self.value [ 652.946114] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.946114] env[61768]: updated_port = self._update_port( [ 652.946114] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.946114] env[61768]: _ensure_no_port_binding_failure(port) [ 652.946114] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.946114] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.946114] env[61768]: nova.exception.PortBindingFailed: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. [ 652.946114] env[61768]: Removing descriptor: 17 [ 652.950558] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397f1e9e-1ba2-4b05-9c1f-6c5019e52653 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.964989] env[61768]: ERROR nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Traceback (most recent call last): [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] yield resources [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self.driver.spawn(context, instance, image_meta, [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] vm_ref = self.build_virtual_machine(instance, [ 652.964989] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] for vif in network_info: [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] return self._sync_wrapper(fn, *args, **kwargs) [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self.wait() [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self[:] = self._gt.wait() [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] return self._exit_event.wait() [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 652.965541] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] current.throw(*self._exc) [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] result = function(*args, **kwargs) [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] return func(*args, **kwargs) [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] raise e [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] nwinfo = self.network_api.allocate_for_instance( [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] created_port_ids = self._update_ports_for_instance( [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] with excutils.save_and_reraise_exception(): [ 652.965982] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self.force_reraise() [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] raise self.value [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] updated_port = self._update_port( [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] _ensure_no_port_binding_failure(port) [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] raise exception.PortBindingFailed(port_id=port['id']) [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] nova.exception.PortBindingFailed: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. [ 652.966330] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] [ 652.966330] env[61768]: INFO nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Terminating instance [ 652.967759] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Acquiring lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.108259] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.331529] env[61768]: DEBUG nova.network.neutron [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.358468] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.358998] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 653.361636] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.568s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.363043] env[61768]: INFO nova.compute.claims [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 653.405130] env[61768]: DEBUG nova.network.neutron [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.871144] env[61768]: DEBUG nova.compute.utils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 653.874646] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 653.874982] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 653.911022] env[61768]: DEBUG oslo_concurrency.lockutils [req-9ac74b78-921e-4353-8e27-b95f5900252d req-a29b4019-8d5c-4ff6-9964-cc9b3bd05e8d service nova] Releasing lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.911022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Acquired lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.911022] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 653.928169] env[61768]: DEBUG nova.policy [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1f5a3ac79854c81830a5186768e5706', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f79d305b334c52a4875f9c3af4ff7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 654.217776] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Successfully created port: 8cea65d9-1b95-49ac-aa19-6b679af78426 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 654.376258] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 654.436414] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 654.572767] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.846289] env[61768]: DEBUG nova.compute.manager [req-858dd041-3ec9-411d-83bd-a27a7af1d4df req-4e363377-3924-46a9-8088-19719d0bb127 service nova] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Received event network-vif-deleted-e9f4b2f1-f86d-4331-8c47-42dbc1299a0e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 654.848484] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ac2f1e-4e3b-4781-9606-6329c1f89c4a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.857276] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aca394a-b2ec-40b8-9d72-8c3948cf4adc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.892475] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee6e3ac-8e9e-4d08-b8bc-9bfe189bca6e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.900825] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7897868-94c7-44f3-8028-7b20fbf4eda5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.917355] env[61768]: DEBUG nova.compute.provider_tree [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.075660] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Releasing lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.079020] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.079020] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 655.079020] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f326478-fd8d-4973-bb1e-57a14c9dca3f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.086994] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80de87cb-c37f-4a8f-a922-467fb005aa96 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.110863] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5 could not be found. [ 655.111111] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 655.111342] env[61768]: INFO nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 655.111555] env[61768]: DEBUG oslo.service.loopingcall [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.111778] env[61768]: DEBUG nova.compute.manager [-] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.112274] env[61768]: DEBUG nova.network.neutron [-] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 655.135473] env[61768]: DEBUG nova.network.neutron [-] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 655.382490] env[61768]: ERROR nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. [ 655.382490] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 655.382490] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.382490] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 655.382490] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.382490] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 655.382490] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.382490] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 655.382490] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.382490] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 655.382490] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.382490] env[61768]: ERROR nova.compute.manager raise self.value [ 655.382490] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.382490] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 655.382490] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.382490] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 655.382980] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.382980] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 655.382980] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. [ 655.382980] env[61768]: ERROR nova.compute.manager [ 655.382980] env[61768]: Traceback (most recent call last): [ 655.382980] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 655.382980] env[61768]: listener.cb(fileno) [ 655.382980] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.382980] env[61768]: result = function(*args, **kwargs) [ 655.382980] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.382980] env[61768]: return func(*args, **kwargs) [ 655.382980] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.382980] env[61768]: raise e [ 655.382980] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.382980] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 655.382980] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.382980] env[61768]: created_port_ids = self._update_ports_for_instance( [ 655.382980] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.382980] env[61768]: with excutils.save_and_reraise_exception(): [ 655.382980] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.382980] env[61768]: self.force_reraise() [ 655.382980] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.382980] env[61768]: raise self.value [ 655.382980] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.382980] env[61768]: updated_port = self._update_port( [ 655.382980] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.382980] env[61768]: _ensure_no_port_binding_failure(port) [ 655.382980] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.382980] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 655.383852] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. [ 655.383852] env[61768]: Removing descriptor: 17 [ 655.397529] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 655.420340] env[61768]: DEBUG nova.scheduler.client.report [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 655.425187] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 655.425420] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 655.425579] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 655.425761] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 655.425925] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 655.426547] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 655.426795] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 655.426985] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 655.427308] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 655.427503] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 655.427684] env[61768]: DEBUG nova.virt.hardware [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 655.428992] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45f4414-1da9-421b-ad25-b3ff1a2a6583 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.436760] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f41daf-33cf-4769-b504-3956200f3f8c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.451704] env[61768]: ERROR nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Traceback (most recent call last): [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] yield resources [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self.driver.spawn(context, instance, image_meta, [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] vm_ref = self.build_virtual_machine(instance, [ 655.451704] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] for vif in network_info: [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] return self._sync_wrapper(fn, *args, **kwargs) [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self.wait() [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self[:] = self._gt.wait() [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] return self._exit_event.wait() [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 655.452138] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] current.throw(*self._exc) [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] result = function(*args, **kwargs) [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] return func(*args, **kwargs) [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] raise e [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] nwinfo = self.network_api.allocate_for_instance( [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] created_port_ids = self._update_ports_for_instance( [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] with excutils.save_and_reraise_exception(): [ 655.452527] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self.force_reraise() [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] raise self.value [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] updated_port = self._update_port( [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] _ensure_no_port_binding_failure(port) [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] raise exception.PortBindingFailed(port_id=port['id']) [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] nova.exception.PortBindingFailed: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. [ 655.452912] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] [ 655.452912] env[61768]: INFO nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Terminating instance [ 655.453732] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.453921] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquired lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.454113] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 655.637458] env[61768]: DEBUG nova.network.neutron [-] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.933066] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.571s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.933680] env[61768]: DEBUG nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.936207] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.218s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.974607] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.067850] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.141050] env[61768]: INFO nova.compute.manager [-] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Took 1.03 seconds to deallocate network for instance. [ 656.144195] env[61768]: DEBUG nova.compute.claims [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.144384] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.445000] env[61768]: DEBUG nova.compute.utils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 656.448481] env[61768]: DEBUG nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Not allocating networking since 'none' was specified. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 656.570699] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Releasing lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.571983] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 656.571983] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 656.571983] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0a2671a1-6376-4714-9b5c-88c1c977f862 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.584335] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41d89128-66e7-4657-95d9-3e6f1a79bd57 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.608837] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7d35cf35-d264-468d-a9ab-7b41dd020023 could not be found. [ 656.609085] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 656.609280] env[61768]: INFO nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Took 0.04 seconds to destroy the instance on the hypervisor. [ 656.609518] env[61768]: DEBUG oslo.service.loopingcall [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 656.611704] env[61768]: DEBUG nova.compute.manager [-] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.611816] env[61768]: DEBUG nova.network.neutron [-] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 656.626260] env[61768]: DEBUG nova.network.neutron [-] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.797030] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcb5559-1014-457d-8647-a7c95340635e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.804005] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dfc2e95-95be-4c03-9e60-4cca49caf08b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.833732] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1746817-f944-4faf-a3b8-3b3aa0b67cf2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.841467] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f6ede6-7ece-48dc-8ea0-74d82cdf9be7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.854766] env[61768]: DEBUG nova.compute.provider_tree [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.869396] env[61768]: DEBUG nova.compute.manager [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Received event network-changed-8cea65d9-1b95-49ac-aa19-6b679af78426 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 656.869671] env[61768]: DEBUG nova.compute.manager [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Refreshing instance network info cache due to event network-changed-8cea65d9-1b95-49ac-aa19-6b679af78426. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 656.869868] env[61768]: DEBUG oslo_concurrency.lockutils [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] Acquiring lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.869934] env[61768]: DEBUG oslo_concurrency.lockutils [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] Acquired lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.870148] env[61768]: DEBUG nova.network.neutron [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Refreshing network info cache for port 8cea65d9-1b95-49ac-aa19-6b679af78426 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 656.948766] env[61768]: DEBUG nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 657.128515] env[61768]: DEBUG nova.network.neutron [-] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.358508] env[61768]: DEBUG nova.scheduler.client.report [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 657.393410] env[61768]: DEBUG nova.network.neutron [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.474018] env[61768]: DEBUG nova.network.neutron [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.630814] env[61768]: INFO nova.compute.manager [-] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Took 1.02 seconds to deallocate network for instance. [ 657.633179] env[61768]: DEBUG nova.compute.claims [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 657.633365] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.863479] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.864154] env[61768]: ERROR nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Traceback (most recent call last): [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self.driver.spawn(context, instance, image_meta, [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] vm_ref = self.build_virtual_machine(instance, [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.864154] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] for vif in network_info: [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] return self._sync_wrapper(fn, *args, **kwargs) [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self.wait() [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self[:] = self._gt.wait() [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] return self._exit_event.wait() [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] current.throw(*self._exc) [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.865048] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] result = function(*args, **kwargs) [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] return func(*args, **kwargs) [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] raise e [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] nwinfo = self.network_api.allocate_for_instance( [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] created_port_ids = self._update_ports_for_instance( [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] with excutils.save_and_reraise_exception(): [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] self.force_reraise() [ 657.865952] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] raise self.value [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] updated_port = self._update_port( [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] _ensure_no_port_binding_failure(port) [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] raise exception.PortBindingFailed(port_id=port['id']) [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] nova.exception.PortBindingFailed: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. [ 657.866552] env[61768]: ERROR nova.compute.manager [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] [ 657.866552] env[61768]: DEBUG nova.compute.utils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.867041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.596s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.867589] env[61768]: INFO nova.compute.claims [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.870148] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Build of instance e6922bb7-c2df-468a-aa5b-d03e274b8ecc was re-scheduled: Binding failed for port 5bd4362c-14c9-4465-bef1-564b38320b12, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.870570] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.870795] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Acquiring lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.870946] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Acquired lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.871124] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.960849] env[61768]: DEBUG nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.976439] env[61768]: DEBUG oslo_concurrency.lockutils [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] Releasing lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.976690] env[61768]: DEBUG nova.compute.manager [req-07a02825-a5dd-49c5-8e6e-39647543c255 req-43d9e543-c82a-49b0-b9a4-669bc8141e30 service nova] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Received event network-vif-deleted-8cea65d9-1b95-49ac-aa19-6b679af78426 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.985225] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.985464] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.985621] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.985805] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.985994] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.986167] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.986382] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.986542] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.986706] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.986867] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.987124] env[61768]: DEBUG nova.virt.hardware [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.987979] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-872c54d4-d2a3-457d-ad25-186c352926dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.996307] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa98d19-f69d-4e88-a434-3a25ac4c06c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.009147] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 658.014613] env[61768]: DEBUG oslo.service.loopingcall [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.014833] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 658.015038] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98bdaca5-49f6-4ce3-9b9e-d1b14b21534f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.030712] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 658.030712] env[61768]: value = "task-1228565" [ 658.030712] env[61768]: _type = "Task" [ 658.030712] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.037846] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228565, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.391260] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.471543] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.540475] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228565, 'name': CreateVM_Task, 'duration_secs': 0.256733} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.541366] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 658.541366] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 658.541487] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.541695] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 658.541933] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab4ab277-b977-41e1-bf72-4556739fda09 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.546193] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 658.546193] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5279e997-b0f1-6950-4f38-600366ed3847" [ 658.546193] env[61768]: _type = "Task" [ 658.546193] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.553178] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5279e997-b0f1-6950-4f38-600366ed3847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.974337] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Releasing lock "refresh_cache-e6922bb7-c2df-468a-aa5b-d03e274b8ecc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.974599] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.974768] env[61768]: DEBUG nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.974937] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.991578] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 659.055579] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5279e997-b0f1-6950-4f38-600366ed3847, 'name': SearchDatastore_Task, 'duration_secs': 0.010246} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.057907] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.058236] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 659.058388] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.058537] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.058715] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 659.059157] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dc7894d-7245-46d4-bccc-f94515f562af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.071447] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 659.071634] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 659.072366] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a4c279e9-995f-4735-b156-71b4803fdeb5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.077391] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 659.077391] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522ab8ca-5693-8a63-1950-d3b39b54b8af" [ 659.077391] env[61768]: _type = "Task" [ 659.077391] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.086689] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522ab8ca-5693-8a63-1950-d3b39b54b8af, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.213786] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efa965f-0529-4a98-bbc4-f836bb032dd0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.222438] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0226ec32-6f77-47b3-b495-080f5fa1071f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.253105] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ab391d-4da5-4681-a94c-232f09e7b00d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.260405] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89533ba-d466-41d9-a00f-48f3f597f434 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.273642] env[61768]: DEBUG nova.compute.provider_tree [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.494337] env[61768]: DEBUG nova.network.neutron [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.587858] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522ab8ca-5693-8a63-1950-d3b39b54b8af, 'name': SearchDatastore_Task, 'duration_secs': 0.00865} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.588503] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e48d2a9-91c2-47aa-920d-97c255430d3d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.593231] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 659.593231] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bfc417-2c1e-795e-9cc4-c27437fddd46" [ 659.593231] env[61768]: _type = "Task" [ 659.593231] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.601843] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bfc417-2c1e-795e-9cc4-c27437fddd46, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.777071] env[61768]: DEBUG nova.scheduler.client.report [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.998038] env[61768]: INFO nova.compute.manager [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] [instance: e6922bb7-c2df-468a-aa5b-d03e274b8ecc] Took 1.02 seconds to deallocate network for instance. [ 660.104692] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bfc417-2c1e-795e-9cc4-c27437fddd46, 'name': SearchDatastore_Task, 'duration_secs': 0.008342} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.104970] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.105247] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 660.105503] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-774bdd40-dca3-4bf1-89bb-21b542b76f81 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.112466] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 660.112466] env[61768]: value = "task-1228566" [ 660.112466] env[61768]: _type = "Task" [ 660.112466] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.122047] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228566, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.283471] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.284135] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 660.287055] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.562s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 660.623676] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228566, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.430761} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.623963] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 660.624217] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 660.624464] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-805f2cd0-f1ef-49df-b9f1-f728cc58bbb1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.630754] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 660.630754] env[61768]: value = "task-1228567" [ 660.630754] env[61768]: _type = "Task" [ 660.630754] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.638504] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228567, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.791732] env[61768]: DEBUG nova.compute.utils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.796425] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.796425] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 660.849627] env[61768]: DEBUG nova.policy [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ed59d2bcee54afab032a65ae3b21023', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ecf64533220b4a5b9f05c0c864e408a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 661.028099] env[61768]: INFO nova.scheduler.client.report [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Deleted allocations for instance e6922bb7-c2df-468a-aa5b-d03e274b8ecc [ 661.141466] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228567, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060379} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.143873] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 661.144994] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d729896-5c86-42d6-be0f-6457c35a8b04 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.164358] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 661.165243] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Successfully created port: 19414399-40b9-45d1-9b9e-4504496f0a9f {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.167411] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eaa858fd-2390-4e74-989c-ec23158e37ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.181570] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43342e7c-b873-43f0-818c-29f22ca80e31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.190419] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9df8181-e5ea-4d5d-893a-c3a80e4f9873 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.193371] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 661.193371] env[61768]: value = "task-1228568" [ 661.193371] env[61768]: _type = "Task" [ 661.193371] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.222301] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fe43f3-4288-4d0d-b91e-f45a220177f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.227889] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228568, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.232599] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb5995f-acb7-4524-a4af-cf4801879ded {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.245303] env[61768]: DEBUG nova.compute.provider_tree [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.296753] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 661.450075] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Successfully created port: cee8783a-07b1-4362-af37-f11135246697 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 661.542332] env[61768]: DEBUG oslo_concurrency.lockutils [None req-87ede2d9-aded-475f-a776-8609073cf1c1 tempest-ServersTestManualDisk-396056078 tempest-ServersTestManualDisk-396056078-project-member] Lock "e6922bb7-c2df-468a-aa5b-d03e274b8ecc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.077s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.703545] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228568, 'name': ReconfigVM_Task, 'duration_secs': 0.26846} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.703920] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 661.704662] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4d20508-e66c-4b54-a7ce-fb5dde1f3288 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.711773] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 661.711773] env[61768]: value = "task-1228569" [ 661.711773] env[61768]: _type = "Task" [ 661.711773] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.720917] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228569, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.748146] env[61768]: DEBUG nova.scheduler.client.report [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.013872] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Successfully created port: 3570ab93-f9e1-4b8a-a620-cc0b3ada0567 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.045374] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 662.226939] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228569, 'name': Rename_Task, 'duration_secs': 0.121138} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.227257] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 662.227510] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-304e6e14-3161-4e8b-b027-aee8f8c3c7ed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.237941] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 662.237941] env[61768]: value = "task-1228570" [ 662.237941] env[61768]: _type = "Task" [ 662.237941] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.245869] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.257851] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.971s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.258873] env[61768]: ERROR nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Traceback (most recent call last): [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self.driver.spawn(context, instance, image_meta, [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] vm_ref = self.build_virtual_machine(instance, [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.258873] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] for vif in network_info: [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] return self._sync_wrapper(fn, *args, **kwargs) [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self.wait() [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self[:] = self._gt.wait() [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] return self._exit_event.wait() [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] current.throw(*self._exc) [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.259201] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] result = function(*args, **kwargs) [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] return func(*args, **kwargs) [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] raise e [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] nwinfo = self.network_api.allocate_for_instance( [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] created_port_ids = self._update_ports_for_instance( [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] with excutils.save_and_reraise_exception(): [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] self.force_reraise() [ 662.259606] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] raise self.value [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] updated_port = self._update_port( [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] _ensure_no_port_binding_failure(port) [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] raise exception.PortBindingFailed(port_id=port['id']) [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] nova.exception.PortBindingFailed: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. [ 662.259951] env[61768]: ERROR nova.compute.manager [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] [ 662.260340] env[61768]: DEBUG nova.compute.utils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 662.261602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.327s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.263155] env[61768]: INFO nova.compute.claims [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.267531] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Build of instance 07d8d6b6-3cc5-48da-9d4d-563a5fe27442 was re-scheduled: Binding failed for port 6b219d2a-ed79-4bfd-b45b-5ea6d8d90073, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 662.268054] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 662.268368] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Acquiring lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.268590] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Acquired lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.268836] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.308461] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 662.339049] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.339049] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.339049] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.339049] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.339369] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.339369] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.339369] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.339369] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.339369] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.339529] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.339529] env[61768]: DEBUG nova.virt.hardware [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.339589] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c5a4de-8f48-40b5-af7b-364beb79a67f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.348220] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c43185-182f-451b-8612-0e9c1005e13f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.571997] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.749945] env[61768]: DEBUG oslo_vmware.api [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228570, 'name': PowerOnVM_Task, 'duration_secs': 0.401498} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.750393] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 662.750701] env[61768]: INFO nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Took 4.79 seconds to spawn the instance on the hypervisor. [ 662.750981] env[61768]: DEBUG nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 662.751879] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b94003-1913-4dd8-a58e-134fd151cebd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.804517] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.939163] env[61768]: DEBUG nova.compute.manager [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Received event network-changed-19414399-40b9-45d1-9b9e-4504496f0a9f {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 662.939379] env[61768]: DEBUG nova.compute.manager [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Refreshing instance network info cache due to event network-changed-19414399-40b9-45d1-9b9e-4504496f0a9f. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 662.939997] env[61768]: DEBUG oslo_concurrency.lockutils [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] Acquiring lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.940190] env[61768]: DEBUG oslo_concurrency.lockutils [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] Acquired lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.940367] env[61768]: DEBUG nova.network.neutron [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Refreshing network info cache for port 19414399-40b9-45d1-9b9e-4504496f0a9f {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 662.961288] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.129096] env[61768]: ERROR nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. [ 663.129096] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 663.129096] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.129096] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 663.129096] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.129096] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 663.129096] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.129096] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 663.129096] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.129096] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 663.129096] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.129096] env[61768]: ERROR nova.compute.manager raise self.value [ 663.129096] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.129096] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 663.129096] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.129096] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 663.129897] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.129897] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 663.129897] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. [ 663.129897] env[61768]: ERROR nova.compute.manager [ 663.129897] env[61768]: Traceback (most recent call last): [ 663.129897] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 663.129897] env[61768]: listener.cb(fileno) [ 663.129897] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.129897] env[61768]: result = function(*args, **kwargs) [ 663.129897] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.129897] env[61768]: return func(*args, **kwargs) [ 663.129897] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.129897] env[61768]: raise e [ 663.129897] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.129897] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 663.129897] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.129897] env[61768]: created_port_ids = self._update_ports_for_instance( [ 663.129897] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.129897] env[61768]: with excutils.save_and_reraise_exception(): [ 663.129897] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.129897] env[61768]: self.force_reraise() [ 663.129897] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.129897] env[61768]: raise self.value [ 663.129897] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.129897] env[61768]: updated_port = self._update_port( [ 663.129897] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.129897] env[61768]: _ensure_no_port_binding_failure(port) [ 663.129897] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.129897] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 663.130758] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. [ 663.130758] env[61768]: Removing descriptor: 17 [ 663.130758] env[61768]: ERROR nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Traceback (most recent call last): [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] yield resources [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self.driver.spawn(context, instance, image_meta, [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.130758] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] vm_ref = self.build_virtual_machine(instance, [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] for vif in network_info: [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return self._sync_wrapper(fn, *args, **kwargs) [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self.wait() [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self[:] = self._gt.wait() [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return self._exit_event.wait() [ 663.131141] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] result = hub.switch() [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return self.greenlet.switch() [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] result = function(*args, **kwargs) [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return func(*args, **kwargs) [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] raise e [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] nwinfo = self.network_api.allocate_for_instance( [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.131568] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] created_port_ids = self._update_ports_for_instance( [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] with excutils.save_and_reraise_exception(): [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self.force_reraise() [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] raise self.value [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] updated_port = self._update_port( [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] _ensure_no_port_binding_failure(port) [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.131941] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] raise exception.PortBindingFailed(port_id=port['id']) [ 663.132311] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] nova.exception.PortBindingFailed: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. [ 663.132311] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] [ 663.132311] env[61768]: INFO nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Terminating instance [ 663.133017] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.271265] env[61768]: INFO nova.compute.manager [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Took 40.50 seconds to build instance. [ 663.470543] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Releasing lock "refresh_cache-07d8d6b6-3cc5-48da-9d4d-563a5fe27442" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.470543] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 663.470543] env[61768]: DEBUG nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.470543] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.500952] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.510906] env[61768]: DEBUG nova.network.neutron [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.633475] env[61768]: DEBUG nova.network.neutron [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.647817] env[61768]: INFO nova.compute.manager [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Rebuilding instance [ 663.683747] env[61768]: DEBUG nova.compute.manager [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 663.684570] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c1fdfa-7cc0-4595-901e-efb60916017d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.721264] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b90ed90-efc0-40b7-9278-079fcf417307 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.729018] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019bffcd-ed48-4d40-8a43-fdf4c8115350 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.761473] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ca98a5-bbec-4dd0-87d3-c6ad23da3a2d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.768422] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5cc72b0-81d8-40d0-a2f4-55c99e5dcde7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.772889] env[61768]: DEBUG oslo_concurrency.lockutils [None req-08c14ec4-58df-4013-85ef-dbadc151d684 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "1eb76ef5-377c-4447-92da-4c61345c0070" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.357s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.783153] env[61768]: DEBUG nova.compute.provider_tree [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.004830] env[61768]: DEBUG nova.network.neutron [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.136653] env[61768]: DEBUG oslo_concurrency.lockutils [req-2bf84fcd-c6b1-4553-88c1-afd648fca120 req-eed6c5de-9628-4e28-8d23-d5525de5beaf service nova] Releasing lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.137174] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquired lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.137387] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 664.197257] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 664.197560] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3711df44-929d-4459-bfc6-47131e45715b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.204213] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 664.204213] env[61768]: value = "task-1228571" [ 664.204213] env[61768]: _type = "Task" [ 664.204213] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.212625] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228571, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.275498] env[61768]: DEBUG nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 664.285794] env[61768]: DEBUG nova.scheduler.client.report [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.507656] env[61768]: INFO nova.compute.manager [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] [instance: 07d8d6b6-3cc5-48da-9d4d-563a5fe27442] Took 1.04 seconds to deallocate network for instance. [ 664.686906] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.718023] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228571, 'name': PowerOffVM_Task, 'duration_secs': 0.110854} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.718023] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 664.718023] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 664.718023] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1adcf6-ae8c-418a-a1e7-eb64c3ab06c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.725236] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 664.725236] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8f37f36-8402-4e8d-9ee4-30d26bd2fd81 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.754023] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 664.754023] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 664.754023] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleting the datastore file [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 664.754023] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4115a656-7957-4b18-891a-84fe12b31e5d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.758269] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 664.758269] env[61768]: value = "task-1228573" [ 664.758269] env[61768]: _type = "Task" [ 664.758269] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.766457] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228573, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.794021] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.529s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.794021] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.794331] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.855s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.796217] env[61768]: INFO nova.compute.claims [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.803438] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.848677] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.967550] env[61768]: DEBUG nova.compute.manager [req-66cc453c-22d4-4fde-99fb-250d85dd0075 req-83765b77-b4b6-4101-bc07-a6c07e8cc918 service nova] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Received event network-vif-deleted-19414399-40b9-45d1-9b9e-4504496f0a9f {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.270183] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228573, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.119517} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.270183] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 665.270183] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 665.270183] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 665.300420] env[61768]: DEBUG nova.compute.utils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 665.307907] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 665.308126] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 665.351456] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Releasing lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.351887] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.352092] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 665.352395] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03d9a1a9-375c-4289-8d66-842efb29d69a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.361659] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ac4e6d-a73c-4692-9dd9-6391eac459bf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.386513] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6035033d-9e22-49bd-a42c-3faeeb7c5d21 could not be found. [ 665.387273] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 665.387273] env[61768]: INFO nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Took 0.03 seconds to destroy the instance on the hypervisor. [ 665.387273] env[61768]: DEBUG oslo.service.loopingcall [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 665.390065] env[61768]: DEBUG nova.policy [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1f5a3ac79854c81830a5186768e5706', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f79d305b334c52a4875f9c3af4ff7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 665.390220] env[61768]: DEBUG nova.compute.manager [-] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 665.390313] env[61768]: DEBUG nova.network.neutron [-] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 665.456224] env[61768]: DEBUG nova.network.neutron [-] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.545015] env[61768]: INFO nova.scheduler.client.report [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Deleted allocations for instance 07d8d6b6-3cc5-48da-9d4d-563a5fe27442 [ 665.808992] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.978072] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Successfully created port: 4c58dbcd-d586-4dca-956b-056d18f397ef {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 666.056491] env[61768]: DEBUG oslo_concurrency.lockutils [None req-66ff996f-3ef8-41d5-8693-195d209560ff tempest-ServersNegativeTestJSON-1979092056 tempest-ServersNegativeTestJSON-1979092056-project-member] Lock "07d8d6b6-3cc5-48da-9d4d-563a5fe27442" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.099s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.237208] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d001db8-a2cc-4998-9bce-7589528a85f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.247301] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f309be-fdab-4797-a00f-5bc6e1afe1d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.280256] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-392149b3-6f46-4a25-8861-2e26eeec3482 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.290533] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53be39e9-002f-43b4-8def-52d775a99960 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.305430] env[61768]: DEBUG nova.compute.provider_tree [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.320168] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.320168] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.320168] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.320168] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.320450] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.320450] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.320450] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.320450] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.320450] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.320604] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.320604] env[61768]: DEBUG nova.virt.hardware [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.323063] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df07f32f-abd3-48e0-a2c7-ad72f8e0f07e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.331034] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05fa794-436f-4de2-9b55-bd10d75e9ead {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.346554] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.353858] env[61768]: DEBUG oslo.service.loopingcall [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.353858] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 666.353858] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62c8e9f9-7756-4e29-bf4e-c68c9bac43ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.373367] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.373367] env[61768]: value = "task-1228574" [ 666.373367] env[61768]: _type = "Task" [ 666.373367] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.381505] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228574, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.560153] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 666.811103] env[61768]: DEBUG nova.scheduler.client.report [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.829891] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.862599] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.863386] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.863386] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.863386] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.863386] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.863727] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.863727] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.863794] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.863908] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.865076] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.865319] env[61768]: DEBUG nova.virt.hardware [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.866262] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcd4a4b-680e-4671-9ec9-9aa3e7efdcb3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.883567] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bb16045-7c45-467c-a343-57068882a7e7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.902091] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228574, 'name': CreateVM_Task, 'duration_secs': 0.247829} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.902492] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 666.903348] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.903510] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.903830] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 666.904094] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbde5b9c-0fea-4533-bff9-d11577209948 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.908964] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 666.908964] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522cb1fb-dab9-3391-970e-3fbdb075ce36" [ 666.908964] env[61768]: _type = "Task" [ 666.908964] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.918399] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522cb1fb-dab9-3391-970e-3fbdb075ce36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.947462] env[61768]: DEBUG nova.network.neutron [-] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.092812] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.236969] env[61768]: DEBUG nova.compute.manager [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Received event network-changed-4c58dbcd-d586-4dca-956b-056d18f397ef {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.236969] env[61768]: DEBUG nova.compute.manager [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Refreshing instance network info cache due to event network-changed-4c58dbcd-d586-4dca-956b-056d18f397ef. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 667.237183] env[61768]: DEBUG oslo_concurrency.lockutils [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] Acquiring lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.237334] env[61768]: DEBUG oslo_concurrency.lockutils [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] Acquired lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.237498] env[61768]: DEBUG nova.network.neutron [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Refreshing network info cache for port 4c58dbcd-d586-4dca-956b-056d18f397ef {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 667.318934] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.525s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.320087] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 667.323204] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.179s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.325934] env[61768]: INFO nova.compute.claims [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.389609] env[61768]: ERROR nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. [ 667.389609] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 667.389609] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.389609] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 667.389609] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.389609] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 667.389609] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.389609] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 667.389609] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.389609] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 667.389609] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.389609] env[61768]: ERROR nova.compute.manager raise self.value [ 667.389609] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.389609] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 667.389609] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.389609] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 667.390271] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.390271] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 667.390271] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. [ 667.390271] env[61768]: ERROR nova.compute.manager [ 667.390271] env[61768]: Traceback (most recent call last): [ 667.390271] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 667.390271] env[61768]: listener.cb(fileno) [ 667.390271] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.390271] env[61768]: result = function(*args, **kwargs) [ 667.390271] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.390271] env[61768]: return func(*args, **kwargs) [ 667.390271] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.390271] env[61768]: raise e [ 667.390271] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.390271] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 667.390271] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.390271] env[61768]: created_port_ids = self._update_ports_for_instance( [ 667.390271] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.390271] env[61768]: with excutils.save_and_reraise_exception(): [ 667.390271] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.390271] env[61768]: self.force_reraise() [ 667.390271] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.390271] env[61768]: raise self.value [ 667.390271] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.390271] env[61768]: updated_port = self._update_port( [ 667.390271] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.390271] env[61768]: _ensure_no_port_binding_failure(port) [ 667.390271] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.390271] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 667.391231] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. [ 667.391231] env[61768]: Removing descriptor: 17 [ 667.391231] env[61768]: ERROR nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Traceback (most recent call last): [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] yield resources [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self.driver.spawn(context, instance, image_meta, [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 667.391231] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] vm_ref = self.build_virtual_machine(instance, [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] vif_infos = vmwarevif.get_vif_info(self._session, [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] for vif in network_info: [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return self._sync_wrapper(fn, *args, **kwargs) [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self.wait() [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self[:] = self._gt.wait() [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return self._exit_event.wait() [ 667.391568] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] result = hub.switch() [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return self.greenlet.switch() [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] result = function(*args, **kwargs) [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return func(*args, **kwargs) [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] raise e [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] nwinfo = self.network_api.allocate_for_instance( [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 667.391896] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] created_port_ids = self._update_ports_for_instance( [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] with excutils.save_and_reraise_exception(): [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self.force_reraise() [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] raise self.value [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] updated_port = self._update_port( [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] _ensure_no_port_binding_failure(port) [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 667.392282] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] raise exception.PortBindingFailed(port_id=port['id']) [ 667.392587] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] nova.exception.PortBindingFailed: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. [ 667.392587] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] [ 667.392587] env[61768]: INFO nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Terminating instance [ 667.394189] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.419898] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522cb1fb-dab9-3391-970e-3fbdb075ce36, 'name': SearchDatastore_Task, 'duration_secs': 0.008726} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.420247] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.420494] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 667.420728] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.420874] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.421068] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 667.421332] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad8ede6c-4dff-4af3-a68e-3c616df73381 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.429478] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 667.429679] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 667.430422] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-359a9681-cddf-4457-b742-b399d60ec053 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.435521] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 667.435521] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5259f150-1e67-f5ca-5bd7-b6a4e368b70a" [ 667.435521] env[61768]: _type = "Task" [ 667.435521] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.443449] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5259f150-1e67-f5ca-5bd7-b6a4e368b70a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.451049] env[61768]: INFO nova.compute.manager [-] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Took 2.06 seconds to deallocate network for instance. [ 667.454247] env[61768]: DEBUG nova.compute.claims [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.454247] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.756606] env[61768]: DEBUG nova.network.neutron [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 667.833826] env[61768]: DEBUG nova.compute.utils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.833826] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.833826] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 667.848100] env[61768]: DEBUG nova.network.neutron [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.890787] env[61768]: DEBUG nova.policy [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd1f5a3ac79854c81830a5186768e5706', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4f79d305b334c52a4875f9c3af4ff7d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.947150] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5259f150-1e67-f5ca-5bd7-b6a4e368b70a, 'name': SearchDatastore_Task, 'duration_secs': 0.007675} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.951183] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9b5aeb8-cfac-42af-81b8-132243bc2552 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.954275] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 667.954275] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272cd74-dee7-cebf-cdb9-b2c8395724da" [ 667.954275] env[61768]: _type = "Task" [ 667.954275] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.967734] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272cd74-dee7-cebf-cdb9-b2c8395724da, 'name': SearchDatastore_Task, 'duration_secs': 0.008942} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.967734] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.968095] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 667.968395] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3cc5dd2c-b3a6-435e-bb1c-5f1d2688afc0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.976110] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 667.976110] env[61768]: value = "task-1228575" [ 667.976110] env[61768]: _type = "Task" [ 667.976110] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.985629] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228575, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.170193] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "76c282d1-ddfe-46dc-aa7b-225708443379" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.170430] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "76c282d1-ddfe-46dc-aa7b-225708443379" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.234332] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Successfully created port: 515c131f-f3a2-4033-a5a5-21feef66d101 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.341021] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 668.349797] env[61768]: DEBUG oslo_concurrency.lockutils [req-43f7799c-2a76-481e-a052-52db4f70cc95 req-a1c4ea67-4ffa-472f-8b63-5cc4bc2c9aaf service nova] Releasing lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 668.350249] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquired lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.350377] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 668.486205] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228575, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.437613} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.486205] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 668.487028] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 668.487028] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54f42194-53fc-4ade-b4ca-b42f389f0f55 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.502585] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 668.502585] env[61768]: value = "task-1228576" [ 668.502585] env[61768]: _type = "Task" [ 668.502585] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.515027] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228576, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.880538] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.883945] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b095049-d0a6-43f7-971a-844d2c5592c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.891735] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6841f4d2-e950-4176-9186-710d9a408bca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.927601] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6c99094-4725-4708-876a-388572dd6b5e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.938816] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb4b9ad-225d-4be7-82a9-348af966e5f5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.950890] env[61768]: DEBUG nova.compute.provider_tree [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.012274] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228576, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101269} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.012759] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 669.014071] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ddb33c-f840-48bd-ac17-bf4b283aef4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.037684] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Reconfiguring VM instance instance-00000021 to attach disk [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 669.038115] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ceb5b1d6-cf68-4791-b8da-2884dc1d870b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.055080] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.062087] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 669.062087] env[61768]: value = "task-1228577" [ 669.062087] env[61768]: _type = "Task" [ 669.062087] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.071626] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228577, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.261882] env[61768]: DEBUG nova.compute.manager [req-fe015a95-ebf9-430c-bf17-7848ee63c5f9 req-3e0d377f-6bc4-4303-b462-3e2bf47545f1 service nova] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Received event network-vif-deleted-4c58dbcd-d586-4dca-956b-056d18f397ef {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.357490] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 669.387399] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 669.387399] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 669.387399] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 669.387659] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 669.387659] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 669.387659] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 669.387659] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 669.387659] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 669.387932] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 669.387932] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 669.387932] env[61768]: DEBUG nova.virt.hardware [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 669.388290] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1390393e-d26f-498e-91a2-0dc5f26ecda8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.396243] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b12838-b51f-4ec8-875e-b057c03ba89c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.454544] env[61768]: DEBUG nova.scheduler.client.report [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.500823] env[61768]: ERROR nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. [ 669.500823] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 669.500823] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.500823] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 669.500823] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.500823] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 669.500823] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.500823] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 669.500823] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.500823] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 669.500823] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.500823] env[61768]: ERROR nova.compute.manager raise self.value [ 669.500823] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.500823] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 669.500823] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.500823] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 669.501283] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.501283] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 669.501283] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. [ 669.501283] env[61768]: ERROR nova.compute.manager [ 669.501283] env[61768]: Traceback (most recent call last): [ 669.501283] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 669.501283] env[61768]: listener.cb(fileno) [ 669.501283] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.501283] env[61768]: result = function(*args, **kwargs) [ 669.501283] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.501283] env[61768]: return func(*args, **kwargs) [ 669.501283] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.501283] env[61768]: raise e [ 669.501283] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.501283] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 669.501283] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.501283] env[61768]: created_port_ids = self._update_ports_for_instance( [ 669.501283] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.501283] env[61768]: with excutils.save_and_reraise_exception(): [ 669.501283] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.501283] env[61768]: self.force_reraise() [ 669.501283] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.501283] env[61768]: raise self.value [ 669.501283] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.501283] env[61768]: updated_port = self._update_port( [ 669.501283] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.501283] env[61768]: _ensure_no_port_binding_failure(port) [ 669.501283] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.501283] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 669.502144] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. [ 669.502144] env[61768]: Removing descriptor: 14 [ 669.502144] env[61768]: ERROR nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Traceback (most recent call last): [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] yield resources [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self.driver.spawn(context, instance, image_meta, [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 669.502144] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] vm_ref = self.build_virtual_machine(instance, [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] for vif in network_info: [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return self._sync_wrapper(fn, *args, **kwargs) [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self.wait() [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self[:] = self._gt.wait() [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return self._exit_event.wait() [ 669.502484] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] result = hub.switch() [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return self.greenlet.switch() [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] result = function(*args, **kwargs) [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return func(*args, **kwargs) [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] raise e [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] nwinfo = self.network_api.allocate_for_instance( [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 669.502866] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] created_port_ids = self._update_ports_for_instance( [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] with excutils.save_and_reraise_exception(): [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self.force_reraise() [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] raise self.value [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] updated_port = self._update_port( [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] _ensure_no_port_binding_failure(port) [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 669.503252] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] raise exception.PortBindingFailed(port_id=port['id']) [ 669.503594] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] nova.exception.PortBindingFailed: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. [ 669.503594] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] [ 669.503594] env[61768]: INFO nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Terminating instance [ 669.503901] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.504110] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquired lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.504294] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.559661] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Releasing lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.560156] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 669.560360] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 669.560643] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b66b28d4-992d-484e-b493-a2e10a59d3d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.571151] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228577, 'name': ReconfigVM_Task, 'duration_secs': 0.304335} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.572198] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Reconfigured VM instance instance-00000021 to attach disk [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070/1eb76ef5-377c-4447-92da-4c61345c0070.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 669.572793] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1bc5c7ac-abe8-4628-87ce-541d041d905b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.576709] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade666cb-15af-403f-aae2-171f83e3b0cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.590719] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 669.590719] env[61768]: value = "task-1228578" [ 669.590719] env[61768]: _type = "Task" [ 669.590719] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.600313] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1def011b-674f-4336-967c-96d26d48aa6c could not be found. [ 669.600557] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 669.600751] env[61768]: INFO nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 669.600985] env[61768]: DEBUG oslo.service.loopingcall [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.601722] env[61768]: DEBUG nova.compute.manager [-] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 669.601839] env[61768]: DEBUG nova.network.neutron [-] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 669.606614] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228578, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.619925] env[61768]: DEBUG nova.network.neutron [-] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.959924] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.637s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.960497] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 669.963271] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.855s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.964920] env[61768]: INFO nova.compute.claims [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.021283] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.100775] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228578, 'name': Rename_Task, 'duration_secs': 0.122629} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.101064] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 670.101320] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e44fe39b-9b89-4a3a-84b2-e0da4a4b4572 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.107184] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 670.107184] env[61768]: value = "task-1228579" [ 670.107184] env[61768]: _type = "Task" [ 670.107184] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.110992] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.114683] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228579, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.121981] env[61768]: DEBUG nova.network.neutron [-] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.469995] env[61768]: DEBUG nova.compute.utils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 670.473384] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 670.473631] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 670.524359] env[61768]: DEBUG nova.policy [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c5cf8ed295ab4af1841e93f1c515a38b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bb0cca23251c460292383b960698b106', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 670.617558] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Releasing lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.617707] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.617879] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 670.618192] env[61768]: DEBUG oslo_vmware.api [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228579, 'name': PowerOnVM_Task, 'duration_secs': 0.416623} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.618456] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ebc9229-6613-4142-85d4-0593a5e4984f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.621401] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 670.621675] env[61768]: DEBUG nova.compute.manager [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 670.622420] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7eeae8f-38c9-465a-91da-8a31bb893948 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.627051] env[61768]: INFO nova.compute.manager [-] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Took 1.03 seconds to deallocate network for instance. [ 670.634009] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39624a27-273b-4a5f-a1f5-5629fa3f508b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.646760] env[61768]: DEBUG nova.compute.claims [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 670.646954] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.660162] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 496c3799-7350-4258-a456-58bd3ba9b4d0 could not be found. [ 670.660162] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 670.660162] env[61768]: INFO nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Took 0.04 seconds to destroy the instance on the hypervisor. [ 670.660391] env[61768]: DEBUG oslo.service.loopingcall [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.661030] env[61768]: DEBUG nova.compute.manager [-] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.661030] env[61768]: DEBUG nova.network.neutron [-] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 670.681136] env[61768]: DEBUG nova.network.neutron [-] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 670.850853] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Successfully created port: 0b92b10b-e9d5-418e-9fc7-5134ca8655e5 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 670.974495] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 671.158734] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.185341] env[61768]: DEBUG nova.network.neutron [-] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.316610] env[61768]: DEBUG nova.compute.manager [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Received event network-changed-515c131f-f3a2-4033-a5a5-21feef66d101 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.316816] env[61768]: DEBUG nova.compute.manager [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Refreshing instance network info cache due to event network-changed-515c131f-f3a2-4033-a5a5-21feef66d101. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 671.317047] env[61768]: DEBUG oslo_concurrency.lockutils [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] Acquiring lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.317278] env[61768]: DEBUG oslo_concurrency.lockutils [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] Acquired lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.317470] env[61768]: DEBUG nova.network.neutron [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Refreshing network info cache for port 515c131f-f3a2-4033-a5a5-21feef66d101 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 671.361447] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a402149-e991-4322-81c6-e5edc8fcb634 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.369238] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e425a88b-98dc-4929-ac6f-028346908269 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.400084] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334242ed-4632-4a26-90c4-a13c051ebd03 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.407762] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52c5e388-1ad7-46d0-838f-7cbb5c94aaa1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.422055] env[61768]: DEBUG nova.compute.provider_tree [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.647019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "1eb76ef5-377c-4447-92da-4c61345c0070" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.647019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "1eb76ef5-377c-4447-92da-4c61345c0070" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.647019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "1eb76ef5-377c-4447-92da-4c61345c0070-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.647019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "1eb76ef5-377c-4447-92da-4c61345c0070-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.647516] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "1eb76ef5-377c-4447-92da-4c61345c0070-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.648664] env[61768]: INFO nova.compute.manager [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Terminating instance [ 671.650362] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "refresh_cache-1eb76ef5-377c-4447-92da-4c61345c0070" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.650517] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "refresh_cache-1eb76ef5-377c-4447-92da-4c61345c0070" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.650996] env[61768]: DEBUG nova.network.neutron [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 671.688239] env[61768]: INFO nova.compute.manager [-] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Took 1.03 seconds to deallocate network for instance. [ 671.690298] env[61768]: DEBUG nova.compute.claims [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.690475] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.836600] env[61768]: DEBUG nova.network.neutron [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 671.898349] env[61768]: ERROR nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. [ 671.898349] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 671.898349] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.898349] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 671.898349] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.898349] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 671.898349] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.898349] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 671.898349] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.898349] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 671.898349] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.898349] env[61768]: ERROR nova.compute.manager raise self.value [ 671.898349] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.898349] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 671.898349] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.898349] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 671.899121] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.899121] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 671.899121] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. [ 671.899121] env[61768]: ERROR nova.compute.manager [ 671.899121] env[61768]: Traceback (most recent call last): [ 671.899121] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 671.899121] env[61768]: listener.cb(fileno) [ 671.899121] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 671.899121] env[61768]: result = function(*args, **kwargs) [ 671.899121] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 671.899121] env[61768]: return func(*args, **kwargs) [ 671.899121] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 671.899121] env[61768]: raise e [ 671.899121] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 671.899121] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 671.899121] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 671.899121] env[61768]: created_port_ids = self._update_ports_for_instance( [ 671.899121] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 671.899121] env[61768]: with excutils.save_and_reraise_exception(): [ 671.899121] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 671.899121] env[61768]: self.force_reraise() [ 671.899121] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 671.899121] env[61768]: raise self.value [ 671.899121] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 671.899121] env[61768]: updated_port = self._update_port( [ 671.899121] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 671.899121] env[61768]: _ensure_no_port_binding_failure(port) [ 671.899121] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 671.899121] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 671.899905] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. [ 671.899905] env[61768]: Removing descriptor: 14 [ 671.925288] env[61768]: DEBUG nova.scheduler.client.report [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.946540] env[61768]: DEBUG nova.network.neutron [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.986112] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 672.011805] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 672.011805] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 672.012056] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 672.012137] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 672.012272] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 672.012423] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 672.012632] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 672.012790] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 672.012958] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 672.013135] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 672.013313] env[61768]: DEBUG nova.virt.hardware [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.014217] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e82f99a-3279-4d34-9c38-8435493574d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.024498] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af5da2c1-fd6f-4cf1-b72e-49fc5cd641e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.037843] env[61768]: ERROR nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Traceback (most recent call last): [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] yield resources [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self.driver.spawn(context, instance, image_meta, [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] vm_ref = self.build_virtual_machine(instance, [ 672.037843] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] for vif in network_info: [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] return self._sync_wrapper(fn, *args, **kwargs) [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self.wait() [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self[:] = self._gt.wait() [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] return self._exit_event.wait() [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.038211] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] current.throw(*self._exc) [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] result = function(*args, **kwargs) [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] return func(*args, **kwargs) [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] raise e [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] nwinfo = self.network_api.allocate_for_instance( [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] created_port_ids = self._update_ports_for_instance( [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] with excutils.save_and_reraise_exception(): [ 672.038532] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self.force_reraise() [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] raise self.value [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] updated_port = self._update_port( [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] _ensure_no_port_binding_failure(port) [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] raise exception.PortBindingFailed(port_id=port['id']) [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] nova.exception.PortBindingFailed: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. [ 672.038851] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] [ 672.038851] env[61768]: INFO nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Terminating instance [ 672.040182] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Acquiring lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.040343] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Acquired lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.040507] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 672.168037] env[61768]: DEBUG nova.network.neutron [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.218372] env[61768]: DEBUG nova.network.neutron [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.430688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.467s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.431415] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 672.434993] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.291s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.449452] env[61768]: DEBUG oslo_concurrency.lockutils [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] Releasing lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.449700] env[61768]: DEBUG nova.compute.manager [req-bd5c0bf7-4508-430e-8899-fc4701673196 req-4997ccbb-fdd9-47f3-bfd1-3a26e3b18e54 service nova] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Received event network-vif-deleted-515c131f-f3a2-4033-a5a5-21feef66d101 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 672.618846] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.699303] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.721372] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "refresh_cache-1eb76ef5-377c-4447-92da-4c61345c0070" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.721796] env[61768]: DEBUG nova.compute.manager [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 672.721990] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 672.722858] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2031375a-ac55-4394-adcd-d08912b9fea0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.731209] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 672.731463] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8deea4f-38b1-4ff9-8f61-ed2cda23de8d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.738010] env[61768]: DEBUG oslo_vmware.api [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 672.738010] env[61768]: value = "task-1228580" [ 672.738010] env[61768]: _type = "Task" [ 672.738010] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.746587] env[61768]: DEBUG oslo_vmware.api [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228580, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.940500] env[61768]: DEBUG nova.compute.utils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 672.942227] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 672.942583] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 672.998137] env[61768]: DEBUG nova.policy [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4e6d9e98e9354a43af0b6a3ec2ed9a6e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '200eb3beb4754911b04c3481e99c636e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 673.203757] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Releasing lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.203757] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 673.203757] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 673.203757] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-98a33d6d-5579-4e4a-94b4-921f9e923197 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.214662] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91f5218e-0386-4188-a4d4-995a92cc5431 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.238975] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aa0833ae-99c7-426d-a972-8183667a48f8 could not be found. [ 673.239077] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 673.239332] env[61768]: INFO nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 673.239687] env[61768]: DEBUG oslo.service.loopingcall [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.242361] env[61768]: DEBUG nova.compute.manager [-] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.242496] env[61768]: DEBUG nova.network.neutron [-] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 673.253493] env[61768]: DEBUG oslo_vmware.api [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228580, 'name': PowerOffVM_Task, 'duration_secs': 0.206013} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.255920] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 673.256188] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 673.256630] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07cfb3a8-e52e-43e4-8403-e1a397e58de6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.267821] env[61768]: DEBUG nova.network.neutron [-] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.284181] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 673.284473] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 673.284777] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleting the datastore file [datastore2] 1eb76ef5-377c-4447-92da-4c61345c0070 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 673.284986] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-689dbb62-1fc8-4a51-b336-2fc6c14922d1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.291468] env[61768]: DEBUG oslo_vmware.api [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 673.291468] env[61768]: value = "task-1228582" [ 673.291468] env[61768]: _type = "Task" [ 673.291468] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.303150] env[61768]: DEBUG oslo_vmware.api [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228582, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.333585] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-381ac823-3b74-4270-ad6e-45549d2b1d9c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.340896] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368d8391-fc98-42cd-afd8-e0b4bd6c7ada {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.371483] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d0fd77-29e1-4798-bbb7-f6681f8d8151 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.374877] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Successfully created port: 24e98493-8510-4e62-8af7-2debc912d065 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.381537] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f683c25-fae2-469d-977c-4e7583c057ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.396565] env[61768]: DEBUG nova.compute.provider_tree [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.428620] env[61768]: DEBUG nova.compute.manager [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Received event network-changed-0b92b10b-e9d5-418e-9fc7-5134ca8655e5 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 673.428620] env[61768]: DEBUG nova.compute.manager [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Refreshing instance network info cache due to event network-changed-0b92b10b-e9d5-418e-9fc7-5134ca8655e5. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 673.428757] env[61768]: DEBUG oslo_concurrency.lockutils [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] Acquiring lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.428801] env[61768]: DEBUG oslo_concurrency.lockutils [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] Acquired lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.428942] env[61768]: DEBUG nova.network.neutron [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Refreshing network info cache for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 673.446330] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 673.771454] env[61768]: DEBUG nova.network.neutron [-] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.804958] env[61768]: DEBUG oslo_vmware.api [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228582, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.092194} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.805417] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 673.805714] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 673.806125] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 673.806578] env[61768]: INFO nova.compute.manager [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Took 1.08 seconds to destroy the instance on the hypervisor. [ 673.809962] env[61768]: DEBUG oslo.service.loopingcall [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.809962] env[61768]: DEBUG nova.compute.manager [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.809962] env[61768]: DEBUG nova.network.neutron [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 673.827625] env[61768]: DEBUG nova.network.neutron [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 673.900639] env[61768]: DEBUG nova.scheduler.client.report [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.960779] env[61768]: DEBUG nova.network.neutron [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 674.116216] env[61768]: DEBUG nova.network.neutron [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.273059] env[61768]: INFO nova.compute.manager [-] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Took 1.03 seconds to deallocate network for instance. [ 674.277860] env[61768]: DEBUG nova.compute.claims [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 674.278157] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.329553] env[61768]: DEBUG nova.network.neutron [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.407022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.971s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.407022] env[61768]: ERROR nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. [ 674.407022] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Traceback (most recent call last): [ 674.407022] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.407022] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self.driver.spawn(context, instance, image_meta, [ 674.407022] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 674.407022] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.407022] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.407022] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] vm_ref = self.build_virtual_machine(instance, [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] for vif in network_info: [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] return self._sync_wrapper(fn, *args, **kwargs) [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self.wait() [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self[:] = self._gt.wait() [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] return self._exit_event.wait() [ 674.407358] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] current.throw(*self._exc) [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] result = function(*args, **kwargs) [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] return func(*args, **kwargs) [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] raise e [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] nwinfo = self.network_api.allocate_for_instance( [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] created_port_ids = self._update_ports_for_instance( [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.407687] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] with excutils.save_and_reraise_exception(): [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] self.force_reraise() [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] raise self.value [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] updated_port = self._update_port( [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] _ensure_no_port_binding_failure(port) [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] raise exception.PortBindingFailed(port_id=port['id']) [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] nova.exception.PortBindingFailed: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. [ 674.408012] env[61768]: ERROR nova.compute.manager [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] [ 674.408326] env[61768]: DEBUG nova.compute.utils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 674.411905] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.776s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.411905] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Build of instance 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5 was re-scheduled: Binding failed for port e9f4b2f1-f86d-4331-8c47-42dbc1299a0e, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 674.412416] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 674.413164] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Acquiring lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.413164] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Acquired lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.413164] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 674.417308] env[61768]: ERROR nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. [ 674.417308] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 674.417308] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.417308] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 674.417308] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.417308] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 674.417308] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.417308] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 674.417308] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.417308] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 674.417308] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.417308] env[61768]: ERROR nova.compute.manager raise self.value [ 674.417308] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.417308] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 674.417308] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.417308] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 674.417797] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.417797] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 674.417797] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. [ 674.417797] env[61768]: ERROR nova.compute.manager [ 674.417797] env[61768]: Traceback (most recent call last): [ 674.417797] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 674.417797] env[61768]: listener.cb(fileno) [ 674.417797] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.417797] env[61768]: result = function(*args, **kwargs) [ 674.417797] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.417797] env[61768]: return func(*args, **kwargs) [ 674.417797] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.417797] env[61768]: raise e [ 674.417797] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.417797] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 674.417797] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.417797] env[61768]: created_port_ids = self._update_ports_for_instance( [ 674.417797] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.417797] env[61768]: with excutils.save_and_reraise_exception(): [ 674.417797] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.417797] env[61768]: self.force_reraise() [ 674.417797] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.417797] env[61768]: raise self.value [ 674.417797] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.417797] env[61768]: updated_port = self._update_port( [ 674.417797] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.417797] env[61768]: _ensure_no_port_binding_failure(port) [ 674.417797] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.417797] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 674.418688] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. [ 674.418688] env[61768]: Removing descriptor: 17 [ 674.455084] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 674.482094] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 674.482360] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 674.482521] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 674.482702] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 674.482851] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 674.482996] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 674.483343] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 674.483515] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 674.483683] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 674.483846] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 674.484055] env[61768]: DEBUG nova.virt.hardware [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.485118] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda16067-9897-4f7b-af0d-90c3179201d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.494650] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72480698-ea75-4cba-b612-1c70b1563926 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.508377] env[61768]: ERROR nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Traceback (most recent call last): [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] yield resources [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self.driver.spawn(context, instance, image_meta, [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] vm_ref = self.build_virtual_machine(instance, [ 674.508377] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] for vif in network_info: [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] return self._sync_wrapper(fn, *args, **kwargs) [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self.wait() [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self[:] = self._gt.wait() [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] return self._exit_event.wait() [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 674.508717] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] current.throw(*self._exc) [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] result = function(*args, **kwargs) [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] return func(*args, **kwargs) [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] raise e [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] nwinfo = self.network_api.allocate_for_instance( [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] created_port_ids = self._update_ports_for_instance( [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] with excutils.save_and_reraise_exception(): [ 674.509064] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self.force_reraise() [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] raise self.value [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] updated_port = self._update_port( [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] _ensure_no_port_binding_failure(port) [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] raise exception.PortBindingFailed(port_id=port['id']) [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] nova.exception.PortBindingFailed: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. [ 674.509411] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] [ 674.509411] env[61768]: INFO nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Terminating instance [ 674.510764] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Acquiring lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.510925] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Acquired lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.511106] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 674.618787] env[61768]: DEBUG oslo_concurrency.lockutils [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] Releasing lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.619062] env[61768]: DEBUG nova.compute.manager [req-2e293a93-2b48-4ecb-8f7d-5b630ef27c9a req-85353c72-9ab4-4b4e-be58-d40da84144af service nova] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Received event network-vif-deleted-0b92b10b-e9d5-418e-9fc7-5134ca8655e5 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 674.831998] env[61768]: INFO nova.compute.manager [-] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Took 1.02 seconds to deallocate network for instance. [ 674.931526] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.035875] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.036990] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.094943] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.284853] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e2c5293-9b76-4e1a-ab48-8fd5cfef6bcb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.293412] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e6dffe-2468-483f-9ba7-e204948edaf0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.324716] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0156eab1-a591-4211-91f1-9b5523167a0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.332406] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e65d0a36-e9fb-4868-b1ed-f5c3a300d04e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.338627] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.346801] env[61768]: DEBUG nova.compute.provider_tree [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 675.463116] env[61768]: DEBUG nova.compute.manager [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Received event network-changed-24e98493-8510-4e62-8af7-2debc912d065 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 675.463166] env[61768]: DEBUG nova.compute.manager [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Refreshing instance network info cache due to event network-changed-24e98493-8510-4e62-8af7-2debc912d065. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 675.463377] env[61768]: DEBUG oslo_concurrency.lockutils [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] Acquiring lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.540029] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Releasing lock "refresh_cache-4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.540029] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 675.540029] env[61768]: DEBUG nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.540230] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.562087] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.597905] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Releasing lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.598097] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 675.598308] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 675.598630] env[61768]: DEBUG oslo_concurrency.lockutils [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] Acquired lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.598802] env[61768]: DEBUG nova.network.neutron [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Refreshing network info cache for port 24e98493-8510-4e62-8af7-2debc912d065 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.599874] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-359223c9-ed90-45fc-90db-902520b6e6da {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.609732] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559aac73-514c-44bf-b4ad-4c048ffe6714 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.633560] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6540f636-fdfd-48aa-a7b8-0bbc231741fe could not be found. [ 675.633837] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 675.633958] env[61768]: INFO nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Took 0.04 seconds to destroy the instance on the hypervisor. [ 675.634234] env[61768]: DEBUG oslo.service.loopingcall [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 675.634445] env[61768]: DEBUG nova.compute.manager [-] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 675.634563] env[61768]: DEBUG nova.network.neutron [-] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 675.650305] env[61768]: DEBUG nova.network.neutron [-] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 675.849503] env[61768]: DEBUG nova.scheduler.client.report [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.065417] env[61768]: DEBUG nova.network.neutron [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.118399] env[61768]: DEBUG nova.network.neutron [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.152485] env[61768]: DEBUG nova.network.neutron [-] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.168916] env[61768]: DEBUG nova.network.neutron [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.355015] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.355742] env[61768]: ERROR nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Traceback (most recent call last): [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self.driver.spawn(context, instance, image_meta, [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self._vmops.spawn(context, instance, image_meta, injected_files, [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] vm_ref = self.build_virtual_machine(instance, [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] vif_infos = vmwarevif.get_vif_info(self._session, [ 676.355742] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] for vif in network_info: [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] return self._sync_wrapper(fn, *args, **kwargs) [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self.wait() [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self[:] = self._gt.wait() [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] return self._exit_event.wait() [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] current.throw(*self._exc) [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 676.356255] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] result = function(*args, **kwargs) [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] return func(*args, **kwargs) [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] raise e [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] nwinfo = self.network_api.allocate_for_instance( [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] created_port_ids = self._update_ports_for_instance( [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] with excutils.save_and_reraise_exception(): [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] self.force_reraise() [ 676.356615] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] raise self.value [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] updated_port = self._update_port( [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] _ensure_no_port_binding_failure(port) [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] raise exception.PortBindingFailed(port_id=port['id']) [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] nova.exception.PortBindingFailed: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. [ 676.356996] env[61768]: ERROR nova.compute.manager [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] [ 676.356996] env[61768]: DEBUG nova.compute.utils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.358054] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.786s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.360085] env[61768]: INFO nova.compute.claims [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.363069] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Build of instance 7d35cf35-d264-468d-a9ab-7b41dd020023 was re-scheduled: Binding failed for port 8cea65d9-1b95-49ac-aa19-6b679af78426, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 676.363568] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 676.363836] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.364034] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquired lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.364224] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.568739] env[61768]: INFO nova.compute.manager [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] [instance: 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5] Took 1.03 seconds to deallocate network for instance. [ 676.658213] env[61768]: INFO nova.compute.manager [-] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Took 1.02 seconds to deallocate network for instance. [ 676.660630] env[61768]: DEBUG nova.compute.claims [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 676.660806] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.671556] env[61768]: DEBUG oslo_concurrency.lockutils [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] Releasing lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.671813] env[61768]: DEBUG nova.compute.manager [req-11e72341-57f6-4d53-a9f8-9bb1c95839e4 req-b197b6cd-d22d-47bd-9b33-759b3a2b5de7 service nova] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Received event network-vif-deleted-24e98493-8510-4e62-8af7-2debc912d065 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 676.884136] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.965316] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.469555] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Releasing lock "refresh_cache-7d35cf35-d264-468d-a9ab-7b41dd020023" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.469555] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 677.469691] env[61768]: DEBUG nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 677.469852] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.484163] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.620951] env[61768]: INFO nova.scheduler.client.report [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Deleted allocations for instance 4c112f97-ddf5-4481-bbf9-3c7861aa3cb5 [ 677.772231] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729c36ea-3e36-4fc5-8ab1-d2ef91940927 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.779360] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bec3a00-1731-4ca9-9722-4ab2983c418c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.809789] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a19a059-1e2b-488d-9d0e-9bb10f39c16e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.816551] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1046d7-2c3a-4fb1-a852-2b66dcb3e813 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.829123] env[61768]: DEBUG nova.compute.provider_tree [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 677.988491] env[61768]: DEBUG nova.network.neutron [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.133243] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e99b53c-9e22-4583-8779-dd915855b481 tempest-AttachInterfacesV270Test-597502253 tempest-AttachInterfacesV270Test-597502253-project-member] Lock "4c112f97-ddf5-4481-bbf9-3c7861aa3cb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.636s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.331999] env[61768]: DEBUG nova.scheduler.client.report [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.491849] env[61768]: INFO nova.compute.manager [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 7d35cf35-d264-468d-a9ab-7b41dd020023] Took 1.02 seconds to deallocate network for instance. [ 678.635700] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 678.837217] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.479s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.837849] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 678.840376] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.037s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.841764] env[61768]: INFO nova.compute.claims [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.158363] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 679.347092] env[61768]: DEBUG nova.compute.utils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 679.350911] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 679.350911] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 679.396282] env[61768]: DEBUG nova.policy [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a095b6d033e649c0a60f572c3893f95d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee67e53b0b544fda9ff343cff7a802f5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 679.528268] env[61768]: INFO nova.scheduler.client.report [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Deleted allocations for instance 7d35cf35-d264-468d-a9ab-7b41dd020023 [ 679.717774] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Successfully created port: 33194713-4bfb-4de3-9448-ed567fb0bde9 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 679.859131] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.038359] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ce4740f-a1a2-458b-818c-2ec4a07d92f5 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "7d35cf35-d264-468d-a9ab-7b41dd020023" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.872s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.244156] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627ba47c-3d3d-430a-932b-28c34bd7f9b8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.252274] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64eaa2dc-8c1b-4445-9fa8-492d25dc4ea0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.289756] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841613e2-9d0b-472d-aded-100cdf3286f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.297616] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d73df6-854c-47c8-949d-a1d1285e3f54 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.311986] env[61768]: DEBUG nova.compute.provider_tree [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.540282] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 680.544713] env[61768]: DEBUG nova.compute.manager [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Received event network-changed-33194713-4bfb-4de3-9448-ed567fb0bde9 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 680.544897] env[61768]: DEBUG nova.compute.manager [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Refreshing instance network info cache due to event network-changed-33194713-4bfb-4de3-9448-ed567fb0bde9. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 680.545127] env[61768]: DEBUG oslo_concurrency.lockutils [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] Acquiring lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.545269] env[61768]: DEBUG oslo_concurrency.lockutils [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] Acquired lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.545434] env[61768]: DEBUG nova.network.neutron [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Refreshing network info cache for port 33194713-4bfb-4de3-9448-ed567fb0bde9 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.735181] env[61768]: ERROR nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. [ 680.735181] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 680.735181] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.735181] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 680.735181] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.735181] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 680.735181] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.735181] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 680.735181] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.735181] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 680.735181] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.735181] env[61768]: ERROR nova.compute.manager raise self.value [ 680.735181] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.735181] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 680.735181] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.735181] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 680.735788] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.735788] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 680.735788] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. [ 680.735788] env[61768]: ERROR nova.compute.manager [ 680.735788] env[61768]: Traceback (most recent call last): [ 680.735788] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 680.735788] env[61768]: listener.cb(fileno) [ 680.735788] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.735788] env[61768]: result = function(*args, **kwargs) [ 680.735788] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.735788] env[61768]: return func(*args, **kwargs) [ 680.735788] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.735788] env[61768]: raise e [ 680.735788] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.735788] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 680.735788] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.735788] env[61768]: created_port_ids = self._update_ports_for_instance( [ 680.735788] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.735788] env[61768]: with excutils.save_and_reraise_exception(): [ 680.735788] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.735788] env[61768]: self.force_reraise() [ 680.735788] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.735788] env[61768]: raise self.value [ 680.735788] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.735788] env[61768]: updated_port = self._update_port( [ 680.735788] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.735788] env[61768]: _ensure_no_port_binding_failure(port) [ 680.735788] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.735788] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 680.736791] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. [ 680.736791] env[61768]: Removing descriptor: 14 [ 680.814942] env[61768]: DEBUG nova.scheduler.client.report [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.865293] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 680.889403] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 680.889691] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 680.889861] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 680.890135] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 680.890298] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 680.890450] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 680.890653] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 680.890810] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 680.890972] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 680.891143] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 680.891313] env[61768]: DEBUG nova.virt.hardware [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 680.892226] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb04da3-5cbf-4e2a-9e43-bb5d3a89b2f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.901377] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8be4687-c326-4cb9-9814-9de3647ec4e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.915670] env[61768]: ERROR nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Traceback (most recent call last): [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] yield resources [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self.driver.spawn(context, instance, image_meta, [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] vm_ref = self.build_virtual_machine(instance, [ 680.915670] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] for vif in network_info: [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] return self._sync_wrapper(fn, *args, **kwargs) [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self.wait() [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self[:] = self._gt.wait() [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] return self._exit_event.wait() [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.916135] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] current.throw(*self._exc) [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] result = function(*args, **kwargs) [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] return func(*args, **kwargs) [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] raise e [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] nwinfo = self.network_api.allocate_for_instance( [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] created_port_ids = self._update_ports_for_instance( [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] with excutils.save_and_reraise_exception(): [ 680.916615] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self.force_reraise() [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] raise self.value [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] updated_port = self._update_port( [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] _ensure_no_port_binding_failure(port) [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] raise exception.PortBindingFailed(port_id=port['id']) [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] nova.exception.PortBindingFailed: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. [ 680.917302] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] [ 680.917302] env[61768]: INFO nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Terminating instance [ 680.918277] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.070661] env[61768]: DEBUG nova.network.neutron [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.073148] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.175380] env[61768]: DEBUG nova.network.neutron [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.320168] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.320628] env[61768]: DEBUG nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 681.323070] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.230s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 681.324556] env[61768]: INFO nova.compute.claims [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.678564] env[61768]: DEBUG oslo_concurrency.lockutils [req-00450d2c-7712-466e-9866-bab14379efa5 req-225dd17f-44cb-4991-b6bc-6405a18c1b8b service nova] Releasing lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.678564] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquired lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.678564] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.828588] env[61768]: DEBUG nova.compute.utils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 681.831752] env[61768]: DEBUG nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Not allocating networking since 'none' was specified. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 682.195874] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.270934] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.332570] env[61768]: DEBUG nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 682.565607] env[61768]: DEBUG nova.compute.manager [req-de81bfee-4b9a-4d7d-8947-3a7b1d324e1a req-2a81f399-1b7e-4dfe-bb5f-7a5c5563e1a2 service nova] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Received event network-vif-deleted-33194713-4bfb-4de3-9448-ed567fb0bde9 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.640765] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09db43bb-63f6-4ba0-a34e-c28c1cc2bb27 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.648347] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f792071a-3758-4552-9d78-ed423427ab30 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.678382] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52327c5-0d38-4401-9445-2b7d665f766b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.685550] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ab5050d-92ae-4a26-b623-cc82fd730d0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.698478] env[61768]: DEBUG nova.compute.provider_tree [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.774069] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Releasing lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.774639] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 682.774859] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 682.775169] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3ef644a5-32af-42e8-8fd4-f4cb7bcae227 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.784237] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6d5a146-dccf-4f26-8679-de99ba129474 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.805298] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5da49801-148d-4b26-a0a2-816b1d5f7994 could not be found. [ 682.805510] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 682.805692] env[61768]: INFO nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Took 0.03 seconds to destroy the instance on the hypervisor. [ 682.805930] env[61768]: DEBUG oslo.service.loopingcall [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 682.806154] env[61768]: DEBUG nova.compute.manager [-] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 682.806272] env[61768]: DEBUG nova.network.neutron [-] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 682.819274] env[61768]: DEBUG nova.network.neutron [-] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.202271] env[61768]: DEBUG nova.scheduler.client.report [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.323653] env[61768]: DEBUG nova.network.neutron [-] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.343655] env[61768]: DEBUG nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 683.372023] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 683.372023] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 683.372023] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 683.372023] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 683.372539] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 683.372539] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 683.372539] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 683.372539] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 683.372539] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 683.372704] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 683.372704] env[61768]: DEBUG nova.virt.hardware [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 683.372768] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c843c0b-1175-40d4-a64b-b22c65673764 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.380718] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d534d4b4-2ae5-4b31-9317-53b735ad6c00 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.394098] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 683.399622] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Creating folder: Project (90d5622c70bc41d48cc9f16d5494919a). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 683.400247] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96a18aa7-9070-48e4-a516-79063b07b841 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.410697] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Created folder: Project (90d5622c70bc41d48cc9f16d5494919a) in parent group-v265360. [ 683.410883] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Creating folder: Instances. Parent ref: group-v265374. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 683.411115] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-126031cf-b078-4369-b73a-6d9a5b101019 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.420429] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Created folder: Instances in parent group-v265374. [ 683.420537] env[61768]: DEBUG oslo.service.loopingcall [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 683.420716] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 683.420902] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e00f014-bb06-4ea7-9e2f-eef0006abe37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.438284] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 683.438284] env[61768]: value = "task-1228585" [ 683.438284] env[61768]: _type = "Task" [ 683.438284] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.445421] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228585, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.707223] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.707791] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 683.710872] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.257s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.826132] env[61768]: INFO nova.compute.manager [-] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Took 1.02 seconds to deallocate network for instance. [ 683.828656] env[61768]: DEBUG nova.compute.claims [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 683.828837] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.948572] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228585, 'name': CreateVM_Task, 'duration_secs': 0.250324} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.948710] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 683.949106] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.949275] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.949591] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 683.949842] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3ced96c-7216-491f-9ef5-b309469a8c51 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.954257] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 683.954257] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c9f55e-3387-8f03-119a-638d6ed4d06d" [ 683.954257] env[61768]: _type = "Task" [ 683.954257] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.961577] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c9f55e-3387-8f03-119a-638d6ed4d06d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.219056] env[61768]: DEBUG nova.compute.utils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 684.220608] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 684.220775] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 684.271398] env[61768]: DEBUG nova.policy [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1f743baaa3c942fab40159eaa69446f8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ac9d98f8a2564204ab3b1f2056cd6265', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 684.463909] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c9f55e-3387-8f03-119a-638d6ed4d06d, 'name': SearchDatastore_Task, 'duration_secs': 0.010928} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.466728] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.466728] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 684.466864] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.466945] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.467131] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 684.467566] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-09818a07-50b1-4f56-b032-98d884734601 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.475849] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 684.476075] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 684.478779] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fbdbbe0-851f-4e25-9707-76d8548fdcc3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.483860] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 684.483860] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f2302f-84bf-e119-5a26-6974ed8f860d" [ 684.483860] env[61768]: _type = "Task" [ 684.483860] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.491146] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f2302f-84bf-e119-5a26-6974ed8f860d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.562598] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55847e2-8f3f-4993-ba53-db0cc65856a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.571055] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35667830-ec67-4c1e-a73c-67d133033812 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.604710] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Successfully created port: 7bb429c1-2068-427b-9bd5-986393fcbb09 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 684.607050] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed49d88-4442-416a-bb40-af21f906c6aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.614567] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd73f14-0800-4529-8ffd-e0a0537832fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.627822] env[61768]: DEBUG nova.compute.provider_tree [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.723610] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 684.995838] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f2302f-84bf-e119-5a26-6974ed8f860d, 'name': SearchDatastore_Task, 'duration_secs': 0.007749} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.996924] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af8effd2-6222-463b-bf53-fd916146ac8c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.003564] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 685.003564] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523f6d19-147a-6662-a94c-5d3b7044a988" [ 685.003564] env[61768]: _type = "Task" [ 685.003564] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.017292] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523f6d19-147a-6662-a94c-5d3b7044a988, 'name': SearchDatastore_Task, 'duration_secs': 0.008903} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.017870] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.018271] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 685.018783] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6090135d-af95-49c3-b49c-d807c2cc7495 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.026077] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 685.026077] env[61768]: value = "task-1228586" [ 685.026077] env[61768]: _type = "Task" [ 685.026077] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.035833] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228586, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.133257] env[61768]: DEBUG nova.scheduler.client.report [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.536840] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228586, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43105} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.537122] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 685.537338] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 685.537938] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8ee8f6b3-4b26-416b-b416-3456648a58d0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.543881] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 685.543881] env[61768]: value = "task-1228587" [ 685.543881] env[61768]: _type = "Task" [ 685.543881] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.552487] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228587, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.603013] env[61768]: DEBUG nova.compute.manager [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Received event network-changed-7bb429c1-2068-427b-9bd5-986393fcbb09 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 685.603281] env[61768]: DEBUG nova.compute.manager [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Refreshing instance network info cache due to event network-changed-7bb429c1-2068-427b-9bd5-986393fcbb09. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 685.603439] env[61768]: DEBUG oslo_concurrency.lockutils [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] Acquiring lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.603577] env[61768]: DEBUG oslo_concurrency.lockutils [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] Acquired lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.603735] env[61768]: DEBUG nova.network.neutron [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Refreshing network info cache for port 7bb429c1-2068-427b-9bd5-986393fcbb09 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 685.637623] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.638693] env[61768]: ERROR nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Traceback (most recent call last): [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self.driver.spawn(context, instance, image_meta, [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] vm_ref = self.build_virtual_machine(instance, [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.638693] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] for vif in network_info: [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return self._sync_wrapper(fn, *args, **kwargs) [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self.wait() [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self[:] = self._gt.wait() [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return self._exit_event.wait() [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] result = hub.switch() [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.639175] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return self.greenlet.switch() [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] result = function(*args, **kwargs) [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] return func(*args, **kwargs) [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] raise e [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] nwinfo = self.network_api.allocate_for_instance( [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] created_port_ids = self._update_ports_for_instance( [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] with excutils.save_and_reraise_exception(): [ 685.639536] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] self.force_reraise() [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] raise self.value [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] updated_port = self._update_port( [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] _ensure_no_port_binding_failure(port) [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] raise exception.PortBindingFailed(port_id=port['id']) [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] nova.exception.PortBindingFailed: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. [ 685.639861] env[61768]: ERROR nova.compute.manager [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] [ 685.640200] env[61768]: DEBUG nova.compute.utils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 685.640513] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.994s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.643298] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Build of instance 6035033d-9e22-49bd-a42c-3faeeb7c5d21 was re-scheduled: Binding failed for port 19414399-40b9-45d1-9b9e-4504496f0a9f, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 685.643750] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 685.644492] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 685.644753] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquired lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.644927] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 685.735350] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 685.756968] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 685.757239] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 685.757427] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 685.757629] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 685.757776] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 685.757923] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 685.758144] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 685.758304] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 685.758471] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 685.758640] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 685.758803] env[61768]: DEBUG nova.virt.hardware [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 685.759699] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6ab497-02d6-4b76-bb1f-f1b9be712750 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.767739] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6ae252-5be2-44a6-8aed-04c4bca53616 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.828721] env[61768]: ERROR nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. [ 685.828721] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 685.828721] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.828721] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 685.828721] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.828721] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 685.828721] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.828721] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 685.828721] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.828721] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 685.828721] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.828721] env[61768]: ERROR nova.compute.manager raise self.value [ 685.828721] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.828721] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 685.828721] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.828721] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 685.829183] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.829183] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 685.829183] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. [ 685.829183] env[61768]: ERROR nova.compute.manager [ 685.829183] env[61768]: Traceback (most recent call last): [ 685.829183] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 685.829183] env[61768]: listener.cb(fileno) [ 685.829183] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.829183] env[61768]: result = function(*args, **kwargs) [ 685.829183] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.829183] env[61768]: return func(*args, **kwargs) [ 685.829183] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.829183] env[61768]: raise e [ 685.829183] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.829183] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 685.829183] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.829183] env[61768]: created_port_ids = self._update_ports_for_instance( [ 685.829183] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.829183] env[61768]: with excutils.save_and_reraise_exception(): [ 685.829183] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.829183] env[61768]: self.force_reraise() [ 685.829183] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.829183] env[61768]: raise self.value [ 685.829183] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.829183] env[61768]: updated_port = self._update_port( [ 685.829183] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.829183] env[61768]: _ensure_no_port_binding_failure(port) [ 685.829183] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.829183] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 685.829896] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. [ 685.829896] env[61768]: Removing descriptor: 14 [ 685.829896] env[61768]: ERROR nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Traceback (most recent call last): [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] yield resources [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self.driver.spawn(context, instance, image_meta, [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 685.829896] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] vm_ref = self.build_virtual_machine(instance, [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] vif_infos = vmwarevif.get_vif_info(self._session, [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] for vif in network_info: [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return self._sync_wrapper(fn, *args, **kwargs) [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self.wait() [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self[:] = self._gt.wait() [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return self._exit_event.wait() [ 685.830215] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] result = hub.switch() [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return self.greenlet.switch() [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] result = function(*args, **kwargs) [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return func(*args, **kwargs) [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] raise e [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] nwinfo = self.network_api.allocate_for_instance( [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 685.830553] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] created_port_ids = self._update_ports_for_instance( [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] with excutils.save_and_reraise_exception(): [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self.force_reraise() [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] raise self.value [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] updated_port = self._update_port( [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] _ensure_no_port_binding_failure(port) [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 685.830900] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] raise exception.PortBindingFailed(port_id=port['id']) [ 685.831211] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] nova.exception.PortBindingFailed: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. [ 685.831211] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] [ 685.831211] env[61768]: INFO nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Terminating instance [ 685.832087] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Acquiring lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.053511] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228587, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.054396} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.053780] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 686.054598] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08e939d-8c3c-4564-a744-2053f302b456 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.073598] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 686.073843] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd9c799d-b2f8-4f6b-af16-2774233d36aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.092633] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 686.092633] env[61768]: value = "task-1228588" [ 686.092633] env[61768]: _type = "Task" [ 686.092633] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.100174] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228588, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.121830] env[61768]: DEBUG nova.network.neutron [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.167554] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 686.304100] env[61768]: DEBUG nova.network.neutron [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.316489] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.517350] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb7bc82-a2fe-4daa-b269-82a6059d188a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.524796] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a377c3-12e9-4c55-8025-3afba8f48606 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.555781] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fd9705-f0dc-4d52-bcba-233418bd05fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.563218] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2af36d-209e-40b6-93ca-75bc59014907 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.576622] env[61768]: DEBUG nova.compute.provider_tree [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.604361] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228588, 'name': ReconfigVM_Task, 'duration_secs': 0.274826} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.604634] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 686.605491] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-916878eb-5058-446c-bca1-b4e037ffd6bc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.611516] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 686.611516] env[61768]: value = "task-1228589" [ 686.611516] env[61768]: _type = "Task" [ 686.611516] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.619670] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228589, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.808859] env[61768]: DEBUG oslo_concurrency.lockutils [req-3686eee6-6767-4fae-89c6-5b0153d77c8c req-33838acc-5689-4a62-9735-4a9a11d9aabb service nova] Releasing lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.809296] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Acquired lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.809483] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 686.817201] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Releasing lock "refresh_cache-6035033d-9e22-49bd-a42c-3faeeb7c5d21" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.817201] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 686.817201] env[61768]: DEBUG nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.817408] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 686.832853] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.082514] env[61768]: DEBUG nova.scheduler.client.report [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.121783] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228589, 'name': Rename_Task, 'duration_secs': 0.135527} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.122164] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 687.122332] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2367d5c6-f474-456d-a579-bdb9c0691278 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.128627] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 687.128627] env[61768]: value = "task-1228590" [ 687.128627] env[61768]: _type = "Task" [ 687.128627] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.136020] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228590, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.328188] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 687.335585] env[61768]: DEBUG nova.network.neutron [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.412834] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.588232] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.948s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.588953] env[61768]: ERROR nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Traceback (most recent call last): [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self.driver.spawn(context, instance, image_meta, [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] vm_ref = self.build_virtual_machine(instance, [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] vif_infos = vmwarevif.get_vif_info(self._session, [ 687.588953] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] for vif in network_info: [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return self._sync_wrapper(fn, *args, **kwargs) [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self.wait() [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self[:] = self._gt.wait() [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return self._exit_event.wait() [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] result = hub.switch() [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 687.589316] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return self.greenlet.switch() [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] result = function(*args, **kwargs) [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] return func(*args, **kwargs) [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] raise e [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] nwinfo = self.network_api.allocate_for_instance( [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] created_port_ids = self._update_ports_for_instance( [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] with excutils.save_and_reraise_exception(): [ 687.589652] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] self.force_reraise() [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] raise self.value [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] updated_port = self._update_port( [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] _ensure_no_port_binding_failure(port) [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] raise exception.PortBindingFailed(port_id=port['id']) [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] nova.exception.PortBindingFailed: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. [ 687.589990] env[61768]: ERROR nova.compute.manager [instance: 1def011b-674f-4336-967c-96d26d48aa6c] [ 687.590351] env[61768]: DEBUG nova.compute.utils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 687.591154] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.432s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.591376] env[61768]: DEBUG nova.objects.instance [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 687.598086] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Build of instance 1def011b-674f-4336-967c-96d26d48aa6c was re-scheduled: Binding failed for port 4c58dbcd-d586-4dca-956b-056d18f397ef, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 687.598086] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 687.598086] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 687.598086] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquired lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.598376] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 687.635852] env[61768]: DEBUG nova.compute.manager [req-f0ff8001-c513-43db-aa12-507ba3168df0 req-91979b71-0fb9-4301-ba1c-373169d3c48c service nova] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Received event network-vif-deleted-7bb429c1-2068-427b-9bd5-986393fcbb09 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 687.642519] env[61768]: DEBUG oslo_vmware.api [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228590, 'name': PowerOnVM_Task, 'duration_secs': 0.421343} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.642793] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 687.643029] env[61768]: INFO nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Took 4.30 seconds to spawn the instance on the hypervisor. [ 687.643239] env[61768]: DEBUG nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 687.644068] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249c484c-a783-42e9-aef0-74c8a9f0fa3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.838383] env[61768]: INFO nova.compute.manager [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 6035033d-9e22-49bd-a42c-3faeeb7c5d21] Took 1.02 seconds to deallocate network for instance. [ 687.915579] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Releasing lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.915986] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 687.916193] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 687.916531] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51c2f26d-392c-403f-8c61-344317341bb9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.925628] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84093f63-3cc7-4e91-8b92-ff1366e85702 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.948565] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c299abc5-7913-4eea-ac0a-02e713c6009c could not be found. [ 687.948565] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 687.948684] env[61768]: INFO nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Took 0.03 seconds to destroy the instance on the hypervisor. [ 687.948835] env[61768]: DEBUG oslo.service.loopingcall [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 687.949074] env[61768]: DEBUG nova.compute.manager [-] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.949168] env[61768]: DEBUG nova.network.neutron [-] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 687.962884] env[61768]: DEBUG nova.network.neutron [-] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.116994] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.163629] env[61768]: INFO nova.compute.manager [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Took 23.38 seconds to build instance. [ 688.197608] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.465132] env[61768]: DEBUG nova.network.neutron [-] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.605146] env[61768]: DEBUG oslo_concurrency.lockutils [None req-429f6b77-b912-4e64-8f16-764bbd796947 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.606938] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.916s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.665234] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2a70dee9-486b-4766-8363-40e01103652f tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.706s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.700603] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Releasing lock "refresh_cache-1def011b-674f-4336-967c-96d26d48aa6c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.700853] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 688.701059] env[61768]: DEBUG nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.701222] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 688.716250] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 688.869693] env[61768]: INFO nova.scheduler.client.report [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Deleted allocations for instance 6035033d-9e22-49bd-a42c-3faeeb7c5d21 [ 688.967651] env[61768]: INFO nova.compute.manager [-] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Took 1.02 seconds to deallocate network for instance. [ 688.970944] env[61768]: DEBUG nova.compute.claims [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 688.971162] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.167836] env[61768]: DEBUG nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.219738] env[61768]: DEBUG nova.network.neutron [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.381681] env[61768]: DEBUG oslo_concurrency.lockutils [None req-23cb0b97-7218-42f5-9ced-4c403eda9ed3 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "6035033d-9e22-49bd-a42c-3faeeb7c5d21" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.184s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.406260] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a90fd73a-ba81-471c-a28d-e97a2b0a2b37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.414099] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8dd6267-ad39-4f31-a391-94140012b157 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.443371] env[61768]: INFO nova.compute.manager [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Rebuilding instance [ 689.446365] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f51efe-ec14-4230-a840-e19621a76ef5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.453293] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1274ecff-cab1-4259-8f21-e91211c997ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.466134] env[61768]: DEBUG nova.compute.provider_tree [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.489347] env[61768]: DEBUG nova.compute.manager [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 689.490186] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3a43fc1-290d-40b0-bca9-69e954627f82 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.688209] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.722323] env[61768]: INFO nova.compute.manager [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 1def011b-674f-4336-967c-96d26d48aa6c] Took 1.02 seconds to deallocate network for instance. [ 689.884669] env[61768]: DEBUG nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 689.971467] env[61768]: DEBUG nova.scheduler.client.report [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.002193] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 690.002819] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7013d76-3a5b-4871-84bf-cbfc688609f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.011741] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 690.011741] env[61768]: value = "task-1228591" [ 690.011741] env[61768]: _type = "Task" [ 690.011741] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.020705] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228591, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.411191] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.475396] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.476112] env[61768]: ERROR nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Traceback (most recent call last): [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self.driver.spawn(context, instance, image_meta, [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] vm_ref = self.build_virtual_machine(instance, [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] vif_infos = vmwarevif.get_vif_info(self._session, [ 690.476112] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] for vif in network_info: [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return self._sync_wrapper(fn, *args, **kwargs) [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self.wait() [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self[:] = self._gt.wait() [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return self._exit_event.wait() [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] result = hub.switch() [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 690.476634] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return self.greenlet.switch() [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] result = function(*args, **kwargs) [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] return func(*args, **kwargs) [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] raise e [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] nwinfo = self.network_api.allocate_for_instance( [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] created_port_ids = self._update_ports_for_instance( [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] with excutils.save_and_reraise_exception(): [ 690.476998] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] self.force_reraise() [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] raise self.value [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] updated_port = self._update_port( [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] _ensure_no_port_binding_failure(port) [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] raise exception.PortBindingFailed(port_id=port['id']) [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] nova.exception.PortBindingFailed: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. [ 690.477490] env[61768]: ERROR nova.compute.manager [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] [ 690.477801] env[61768]: DEBUG nova.compute.utils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 690.478243] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.200s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.481730] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Build of instance 496c3799-7350-4258-a456-58bd3ba9b4d0 was re-scheduled: Binding failed for port 515c131f-f3a2-4033-a5a5-21feef66d101, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 690.481864] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 690.482105] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquiring lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.482515] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Acquired lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.482707] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 690.521492] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228591, 'name': PowerOffVM_Task, 'duration_secs': 0.130333} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.522375] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 690.522375] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 690.522965] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a6e2c2-c949-4b1e-a34f-ef6bf9008c49 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.529778] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 690.530614] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d6f38ea8-1e80-490e-9ae1-73bb3de247f7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.556526] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 690.556836] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 690.557078] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Deleting the datastore file [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 690.557388] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db1d23c3-55ce-4e7f-ba6f-6d619933fd4f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.563696] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 690.563696] env[61768]: value = "task-1228593" [ 690.563696] env[61768]: _type = "Task" [ 690.563696] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.573658] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228593, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.753607] env[61768]: INFO nova.scheduler.client.report [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Deleted allocations for instance 1def011b-674f-4336-967c-96d26d48aa6c [ 691.011627] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.073731] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228593, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095801} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.076194] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 691.076415] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 691.077043] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 691.149415] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.263809] env[61768]: DEBUG oslo_concurrency.lockutils [None req-47befa3a-2cc7-47e9-a979-049f05d63154 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "1def011b-674f-4336-967c-96d26d48aa6c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.354s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.335739] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39799190-61a2-4ee5-8814-15f13d738e4e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.346030] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f709caaa-cba4-4ba5-bbe3-0572985bf741 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.373687] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eef5f2c-88c1-48e9-8249-11db489d5f03 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.381430] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a95b4b-fcda-4a1f-a2eb-928973904f95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.394126] env[61768]: DEBUG nova.compute.provider_tree [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 691.654696] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Releasing lock "refresh_cache-496c3799-7350-4258-a456-58bd3ba9b4d0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.654975] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 691.655205] env[61768]: DEBUG nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 691.655382] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 691.672728] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.766243] env[61768]: DEBUG nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 691.897622] env[61768]: DEBUG nova.scheduler.client.report [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 692.116429] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.116776] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.116936] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.117181] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.117372] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.117557] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.117799] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.117961] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.118184] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.118428] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.118650] env[61768]: DEBUG nova.virt.hardware [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.119550] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e252a9-4ef7-4637-b29c-5fb689016b08 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.129143] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcea73d6-8d29-49b4-a56d-8123de3e22e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.142804] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 692.149212] env[61768]: DEBUG oslo.service.loopingcall [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.149634] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 692.149934] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-789a9017-88e2-4bc0-ab85-be90097c98a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.169592] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 692.169592] env[61768]: value = "task-1228594" [ 692.169592] env[61768]: _type = "Task" [ 692.169592] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.177764] env[61768]: DEBUG nova.network.neutron [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.178685] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228594, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.297032] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.403196] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.925s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.404157] env[61768]: ERROR nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Traceback (most recent call last): [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self.driver.spawn(context, instance, image_meta, [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] vm_ref = self.build_virtual_machine(instance, [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.404157] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] for vif in network_info: [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] return self._sync_wrapper(fn, *args, **kwargs) [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self.wait() [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self[:] = self._gt.wait() [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] return self._exit_event.wait() [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] current.throw(*self._exc) [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.405346] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] result = function(*args, **kwargs) [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] return func(*args, **kwargs) [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] raise e [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] nwinfo = self.network_api.allocate_for_instance( [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] created_port_ids = self._update_ports_for_instance( [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] with excutils.save_and_reraise_exception(): [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] self.force_reraise() [ 692.405762] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] raise self.value [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] updated_port = self._update_port( [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] _ensure_no_port_binding_failure(port) [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] raise exception.PortBindingFailed(port_id=port['id']) [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] nova.exception.PortBindingFailed: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. [ 692.406168] env[61768]: ERROR nova.compute.manager [instance: aa0833ae-99c7-426d-a972-8183667a48f8] [ 692.406168] env[61768]: DEBUG nova.compute.utils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 692.407205] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.069s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.407612] env[61768]: DEBUG nova.objects.instance [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lazy-loading 'resources' on Instance uuid 1eb76ef5-377c-4447-92da-4c61345c0070 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 692.413020] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Build of instance aa0833ae-99c7-426d-a972-8183667a48f8 was re-scheduled: Binding failed for port 0b92b10b-e9d5-418e-9fc7-5134ca8655e5, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 692.413020] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 692.413020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Acquiring lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.413020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Acquired lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.413357] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 692.428857] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "77319462-e447-405b-9269-82581effe005" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.429103] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "77319462-e447-405b-9269-82581effe005" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.680073] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228594, 'name': CreateVM_Task, 'duration_secs': 0.251718} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.680656] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 692.680656] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.680832] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.681139] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 692.681618] env[61768]: INFO nova.compute.manager [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] [instance: 496c3799-7350-4258-a456-58bd3ba9b4d0] Took 1.03 seconds to deallocate network for instance. [ 692.683990] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2aacb28c-5bcb-4305-987d-10460c9fa1de {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.689277] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 692.689277] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c71bf8-50a4-8de6-d225-3f7744144ef4" [ 692.689277] env[61768]: _type = "Task" [ 692.689277] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.697034] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c71bf8-50a4-8de6-d225-3f7744144ef4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.939070] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.033964] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.201909] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c71bf8-50a4-8de6-d225-3f7744144ef4, 'name': SearchDatastore_Task, 'duration_secs': 0.008749} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.204613] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.204840] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 693.205102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.205279] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.205464] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 693.205924] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e8dcecd-6fd1-4288-aec2-01613a5f317a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.213934] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 693.214155] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 693.217207] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-643951cb-edb3-497c-89fe-9f247d0d22df {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.224829] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 693.224829] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527fb312-a20f-7bea-3049-6d67030a5a49" [ 693.224829] env[61768]: _type = "Task" [ 693.224829] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.230563] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527fb312-a20f-7bea-3049-6d67030a5a49, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.232065] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fd41143-4ce7-4001-8f2c-e5f6cc3895ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.238470] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246c77a6-abba-44e5-9f53-b0ecd8053b4e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.270397] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3a287d-36a3-4667-8774-275e2e69fbb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.277680] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744e1240-fbf7-4fc5-9a15-44a88b0df872 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.291874] env[61768]: DEBUG nova.compute.provider_tree [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 693.536747] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Releasing lock "refresh_cache-aa0833ae-99c7-426d-a972-8183667a48f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.536963] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 693.537166] env[61768]: DEBUG nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.537338] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 693.552705] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.712107] env[61768]: INFO nova.scheduler.client.report [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Deleted allocations for instance 496c3799-7350-4258-a456-58bd3ba9b4d0 [ 693.734294] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527fb312-a20f-7bea-3049-6d67030a5a49, 'name': SearchDatastore_Task, 'duration_secs': 0.008532} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.735612] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a035ff6-9a3b-4841-b1c8-4f8f60a07679 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.741145] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 693.741145] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ee6946-2464-cf3e-303d-ff079129747d" [ 693.741145] env[61768]: _type = "Task" [ 693.741145] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.749491] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ee6946-2464-cf3e-303d-ff079129747d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.797287] env[61768]: DEBUG nova.scheduler.client.report [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 694.055966] env[61768]: DEBUG nova.network.neutron [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.220713] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f4f5ed78-4986-4e9e-b3e5-43b23e1c1e78 tempest-ListServerFiltersTestJSON-1968034780 tempest-ListServerFiltersTestJSON-1968034780-project-member] Lock "496c3799-7350-4258-a456-58bd3ba9b4d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 150.399s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.251965] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ee6946-2464-cf3e-303d-ff079129747d, 'name': SearchDatastore_Task, 'duration_secs': 0.008327} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.252350] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.252807] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 694.253120] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2df5f6bf-7c62-4fdd-bf99-3acd4ad8d726 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.260591] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 694.260591] env[61768]: value = "task-1228595" [ 694.260591] env[61768]: _type = "Task" [ 694.260591] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.270168] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228595, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.304730] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.897s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.307638] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.647s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.331899] env[61768]: INFO nova.scheduler.client.report [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleted allocations for instance 1eb76ef5-377c-4447-92da-4c61345c0070 [ 694.559399] env[61768]: INFO nova.compute.manager [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] [instance: aa0833ae-99c7-426d-a972-8183667a48f8] Took 1.02 seconds to deallocate network for instance. [ 694.724112] env[61768]: DEBUG nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.773392] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228595, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.425606} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.773662] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 694.773856] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 694.774111] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d8fb5963-c259-4c3a-b0d9-cc56cdc89bc3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.782106] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 694.782106] env[61768]: value = "task-1228596" [ 694.782106] env[61768]: _type = "Task" [ 694.782106] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.791320] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228596, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.839101] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8db6ae96-cdaa-40a2-8cc2-e790542dcccb tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "1eb76ef5-377c-4447-92da-4c61345c0070" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.193s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.129980] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da043831-b11f-4732-a20e-b6a8ba27660e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.138288] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69983846-c15e-47fa-a26b-07188f187670 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.168955] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50c7243f-4550-4e2e-8dce-02d96052550d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.177598] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4662cf30-9e33-425d-b6cd-f8b5fa81645b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.190800] env[61768]: DEBUG nova.compute.provider_tree [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.243289] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.292284] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228596, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055439} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.292544] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 695.294051] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122573bc-c503-4bf1-9a6c-bc1d059fb460 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.313519] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Reconfiguring VM instance instance-00000028 to attach disk [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.313803] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1588f74-3a04-4eb8-8cf5-8b1f3e4b4921 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.333855] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 695.333855] env[61768]: value = "task-1228597" [ 695.333855] env[61768]: _type = "Task" [ 695.333855] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.343483] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228597, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.408121] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "f81aabf6-ae46-405c-b101-12ca707a0567" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.408606] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "f81aabf6-ae46-405c-b101-12ca707a0567" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.408686] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "f81aabf6-ae46-405c-b101-12ca707a0567-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 695.408830] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "f81aabf6-ae46-405c-b101-12ca707a0567-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.409012] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "f81aabf6-ae46-405c-b101-12ca707a0567-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.411673] env[61768]: INFO nova.compute.manager [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Terminating instance [ 695.415512] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "refresh_cache-f81aabf6-ae46-405c-b101-12ca707a0567" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.415897] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquired lock "refresh_cache-f81aabf6-ae46-405c-b101-12ca707a0567" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.415897] env[61768]: DEBUG nova.network.neutron [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.589050] env[61768]: INFO nova.scheduler.client.report [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Deleted allocations for instance aa0833ae-99c7-426d-a972-8183667a48f8 [ 695.693748] env[61768]: DEBUG nova.scheduler.client.report [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.843501] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228597, 'name': ReconfigVM_Task, 'duration_secs': 0.252001} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.843789] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Reconfigured VM instance instance-00000028 to attach disk [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d/63704075-5eaf-4f84-a90b-3a0a3e904a9d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 695.844443] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-25d1630f-3d18-4bae-95db-84c3ee779ec0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.851336] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 695.851336] env[61768]: value = "task-1228598" [ 695.851336] env[61768]: _type = "Task" [ 695.851336] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.858716] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228598, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.938690] env[61768]: DEBUG nova.network.neutron [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.999112] env[61768]: DEBUG nova.network.neutron [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.098952] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4a9dbb9d-ee8c-4f3f-8b87-0fb695c4a9c0 tempest-ServerPasswordTestJSON-2064263980 tempest-ServerPasswordTestJSON-2064263980-project-member] Lock "aa0833ae-99c7-426d-a972-8183667a48f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.781s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.199650] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.892s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.200325] env[61768]: ERROR nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Traceback (most recent call last): [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self.driver.spawn(context, instance, image_meta, [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] vm_ref = self.build_virtual_machine(instance, [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 696.200325] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] for vif in network_info: [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] return self._sync_wrapper(fn, *args, **kwargs) [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self.wait() [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self[:] = self._gt.wait() [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] return self._exit_event.wait() [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] current.throw(*self._exc) [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 696.202175] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] result = function(*args, **kwargs) [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] return func(*args, **kwargs) [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] raise e [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] nwinfo = self.network_api.allocate_for_instance( [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] created_port_ids = self._update_ports_for_instance( [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] with excutils.save_and_reraise_exception(): [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] self.force_reraise() [ 696.202636] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] raise self.value [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] updated_port = self._update_port( [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] _ensure_no_port_binding_failure(port) [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] raise exception.PortBindingFailed(port_id=port['id']) [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] nova.exception.PortBindingFailed: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. [ 696.203060] env[61768]: ERROR nova.compute.manager [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] [ 696.203060] env[61768]: DEBUG nova.compute.utils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 696.203374] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.044s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.203953] env[61768]: INFO nova.compute.claims [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.206938] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Build of instance 6540f636-fdfd-48aa-a7b8-0bbc231741fe was re-scheduled: Binding failed for port 24e98493-8510-4e62-8af7-2debc912d065, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 696.207393] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 696.207637] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Acquiring lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.207775] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Acquired lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.207933] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 696.360843] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228598, 'name': Rename_Task, 'duration_secs': 0.131903} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.361068] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 696.361319] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3b890d0d-7580-45e6-bb32-eae966b6ba12 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.367708] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 696.367708] env[61768]: value = "task-1228599" [ 696.367708] env[61768]: _type = "Task" [ 696.367708] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.375521] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228599, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.501745] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Releasing lock "refresh_cache-f81aabf6-ae46-405c-b101-12ca707a0567" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.505020] env[61768]: DEBUG nova.compute.manager [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 696.505020] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 696.505020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f3274e-3e46-4fb0-837e-15d29f868c7c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.513021] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 696.513021] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-793e0069-b451-4eb8-83d7-b2cbc79c3dc8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.518055] env[61768]: DEBUG oslo_vmware.api [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 696.518055] env[61768]: value = "task-1228600" [ 696.518055] env[61768]: _type = "Task" [ 696.518055] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.525351] env[61768]: DEBUG oslo_vmware.api [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.603329] env[61768]: DEBUG nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.743506] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.882012] env[61768]: DEBUG oslo_vmware.api [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228599, 'name': PowerOnVM_Task, 'duration_secs': 0.420473} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.882384] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 696.882655] env[61768]: DEBUG nova.compute.manager [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 696.883681] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f2e0b2-5e30-437a-b3ca-bfea9158231a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.011024] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.027656] env[61768]: DEBUG oslo_vmware.api [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228600, 'name': PowerOffVM_Task, 'duration_secs': 0.148874} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.028017] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 697.028256] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 697.028547] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-23a53666-45eb-43b1-96ae-5395280aaae9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.052010] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 697.052246] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 697.052430] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleting the datastore file [datastore1] f81aabf6-ae46-405c-b101-12ca707a0567 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 697.052679] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-af34a940-8f0d-46a3-8966-df83e924d995 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.058164] env[61768]: DEBUG oslo_vmware.api [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for the task: (returnval){ [ 697.058164] env[61768]: value = "task-1228602" [ 697.058164] env[61768]: _type = "Task" [ 697.058164] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.066374] env[61768]: DEBUG oslo_vmware.api [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228602, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.131370] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.401459] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.515578] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Releasing lock "refresh_cache-6540f636-fdfd-48aa-a7b8-0bbc231741fe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 697.515881] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 697.516155] env[61768]: DEBUG nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.516437] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.557236] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 697.564023] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00095e10-feb7-4801-893c-57ce179a4874 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.575842] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c342736c-7d55-498b-8d75-9f50c6422d85 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.580832] env[61768]: DEBUG oslo_vmware.api [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Task: {'id': task-1228602, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.446748} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.581164] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 697.584012] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 697.584012] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 697.584012] env[61768]: INFO nova.compute.manager [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Took 1.08 seconds to destroy the instance on the hypervisor. [ 697.584012] env[61768]: DEBUG oslo.service.loopingcall [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 697.584012] env[61768]: DEBUG nova.compute.manager [-] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 697.584012] env[61768]: DEBUG nova.network.neutron [-] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 697.614321] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf2897b-b941-4ad3-a292-50c4d21c80ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.624225] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81ea8aaa-9fd7-401c-8a08-400e5f6afe89 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.642550] env[61768]: DEBUG nova.compute.provider_tree [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.648592] env[61768]: DEBUG nova.network.neutron [-] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.065314] env[61768]: DEBUG nova.network.neutron [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.153782] env[61768]: DEBUG nova.scheduler.client.report [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.159614] env[61768]: DEBUG nova.network.neutron [-] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.161984] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.161984] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.162851] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.163080] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.163250] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.167052] env[61768]: INFO nova.compute.manager [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Terminating instance [ 698.172444] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "refresh_cache-63704075-5eaf-4f84-a90b-3a0a3e904a9d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.172599] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquired lock "refresh_cache-63704075-5eaf-4f84-a90b-3a0a3e904a9d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.172769] env[61768]: DEBUG nova.network.neutron [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 698.570361] env[61768]: INFO nova.compute.manager [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] [instance: 6540f636-fdfd-48aa-a7b8-0bbc231741fe] Took 1.05 seconds to deallocate network for instance. [ 698.660720] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.458s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.660964] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 698.664023] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.590s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.668516] env[61768]: INFO nova.compute.claims [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.671050] env[61768]: INFO nova.compute.manager [-] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Took 1.09 seconds to deallocate network for instance. [ 698.699791] env[61768]: DEBUG nova.network.neutron [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.781290] env[61768]: DEBUG nova.network.neutron [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.174042] env[61768]: DEBUG nova.compute.utils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.180683] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 699.180872] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 699.188203] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.279371] env[61768]: DEBUG nova.policy [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6860fe233824277b6ce262e0ad41120', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'bc72064298d2436b82c16a976bb704e6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.285669] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Releasing lock "refresh_cache-63704075-5eaf-4f84-a90b-3a0a3e904a9d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.285669] env[61768]: DEBUG nova.compute.manager [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 699.285669] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 699.287154] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8de7ca-0f71-4c2d-8691-cd64c23ab16f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.297680] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 699.298014] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-918dcc0d-8925-404e-b5a5-8014baffb48d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.307230] env[61768]: DEBUG oslo_vmware.api [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 699.307230] env[61768]: value = "task-1228603" [ 699.307230] env[61768]: _type = "Task" [ 699.307230] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.320187] env[61768]: DEBUG oslo_vmware.api [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228603, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.609815] env[61768]: INFO nova.scheduler.client.report [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Deleted allocations for instance 6540f636-fdfd-48aa-a7b8-0bbc231741fe [ 699.667612] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Successfully created port: 729d32df-bf2d-4fc2-ae66-57fca1610887 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.681512] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 699.821123] env[61768]: DEBUG oslo_vmware.api [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228603, 'name': PowerOffVM_Task, 'duration_secs': 0.137481} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.821445] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 699.821642] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 699.823027] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b54be25-69e4-41c3-8931-1d368413c079 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.844557] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 699.844792] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 699.845035] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Deleting the datastore file [datastore2] 63704075-5eaf-4f84-a90b-3a0a3e904a9d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 699.845340] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19aacbe1-6529-484c-94d0-1a2cebae57e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.850999] env[61768]: DEBUG oslo_vmware.api [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for the task: (returnval){ [ 699.850999] env[61768]: value = "task-1228605" [ 699.850999] env[61768]: _type = "Task" [ 699.850999] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.858637] env[61768]: DEBUG oslo_vmware.api [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228605, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.069020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d779d69-5dce-494d-b741-24e733a49913 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.076174] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f04b08f-355e-40bf-bff7-7c66b840f665 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.109515] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e2684aa-6188-4be9-9987-7f38c36a7d7a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.118108] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afedcc67-67eb-4549-a34f-40b823f69f88 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.121461] env[61768]: DEBUG oslo_concurrency.lockutils [None req-560efeff-23ff-4776-b54d-0b1a85abe2eb tempest-TenantUsagesTestJSON-889746690 tempest-TenantUsagesTestJSON-889746690-project-member] Lock "6540f636-fdfd-48aa-a7b8-0bbc231741fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.484s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.133097] env[61768]: DEBUG nova.compute.provider_tree [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.364827] env[61768]: DEBUG oslo_vmware.api [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Task: {'id': task-1228605, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.136181} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.364827] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 700.364827] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 700.364827] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 700.365528] env[61768]: INFO nova.compute.manager [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Took 1.08 seconds to destroy the instance on the hypervisor. [ 700.365528] env[61768]: DEBUG oslo.service.loopingcall [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.366778] env[61768]: DEBUG nova.compute.manager [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 700.366961] env[61768]: DEBUG nova.network.neutron [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 700.386420] env[61768]: DEBUG nova.network.neutron [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.554578] env[61768]: DEBUG nova.compute.manager [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Received event network-changed-729d32df-bf2d-4fc2-ae66-57fca1610887 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.554937] env[61768]: DEBUG nova.compute.manager [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Refreshing instance network info cache due to event network-changed-729d32df-bf2d-4fc2-ae66-57fca1610887. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 700.555169] env[61768]: DEBUG oslo_concurrency.lockutils [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] Acquiring lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.555306] env[61768]: DEBUG oslo_concurrency.lockutils [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] Acquired lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.555457] env[61768]: DEBUG nova.network.neutron [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Refreshing network info cache for port 729d32df-bf2d-4fc2-ae66-57fca1610887 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 700.624451] env[61768]: DEBUG nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 700.636068] env[61768]: DEBUG nova.scheduler.client.report [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.695695] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.723083] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.723242] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.723403] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.723584] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.723731] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.723878] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.724146] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.724333] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.724511] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.724713] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.724895] env[61768]: DEBUG nova.virt.hardware [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.725789] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f261325-bbd5-4e8a-978e-c3e0dd8692ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.734396] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e635db-499b-485e-9f7a-12238b346876 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.771891] env[61768]: ERROR nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. [ 700.771891] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 700.771891] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.771891] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 700.771891] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 700.771891] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 700.771891] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 700.771891] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 700.771891] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.771891] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 700.771891] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.771891] env[61768]: ERROR nova.compute.manager raise self.value [ 700.771891] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 700.771891] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 700.771891] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.771891] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 700.772383] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.772383] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 700.772383] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. [ 700.772383] env[61768]: ERROR nova.compute.manager [ 700.772383] env[61768]: Traceback (most recent call last): [ 700.772525] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 700.772525] env[61768]: listener.cb(fileno) [ 700.772525] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.772525] env[61768]: result = function(*args, **kwargs) [ 700.772525] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 700.772525] env[61768]: return func(*args, **kwargs) [ 700.772525] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.772525] env[61768]: raise e [ 700.772525] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.772525] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 700.772525] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 700.772525] env[61768]: created_port_ids = self._update_ports_for_instance( [ 700.772525] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 700.772525] env[61768]: with excutils.save_and_reraise_exception(): [ 700.772525] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.772525] env[61768]: self.force_reraise() [ 700.772525] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.772525] env[61768]: raise self.value [ 700.772525] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 700.772525] env[61768]: updated_port = self._update_port( [ 700.772525] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.772525] env[61768]: _ensure_no_port_binding_failure(port) [ 700.772525] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.772525] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 700.772525] env[61768]: nova.exception.PortBindingFailed: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. [ 700.772525] env[61768]: Removing descriptor: 17 [ 700.773601] env[61768]: ERROR nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Traceback (most recent call last): [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] yield resources [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self.driver.spawn(context, instance, image_meta, [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] vm_ref = self.build_virtual_machine(instance, [ 700.773601] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] vif_infos = vmwarevif.get_vif_info(self._session, [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] for vif in network_info: [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return self._sync_wrapper(fn, *args, **kwargs) [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self.wait() [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self[:] = self._gt.wait() [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return self._exit_event.wait() [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 700.773918] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] result = hub.switch() [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return self.greenlet.switch() [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] result = function(*args, **kwargs) [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return func(*args, **kwargs) [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] raise e [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] nwinfo = self.network_api.allocate_for_instance( [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] created_port_ids = self._update_ports_for_instance( [ 700.774406] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] with excutils.save_and_reraise_exception(): [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self.force_reraise() [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] raise self.value [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] updated_port = self._update_port( [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] _ensure_no_port_binding_failure(port) [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] raise exception.PortBindingFailed(port_id=port['id']) [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] nova.exception.PortBindingFailed: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. [ 700.774792] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] [ 700.775261] env[61768]: INFO nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Terminating instance [ 700.778072] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Acquiring lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 700.890550] env[61768]: DEBUG nova.network.neutron [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.086675] env[61768]: DEBUG nova.network.neutron [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.147922] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.148485] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.152569] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.323s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.159715] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.160933] env[61768]: DEBUG nova.network.neutron [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.397845] env[61768]: INFO nova.compute.manager [-] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Took 1.03 seconds to deallocate network for instance. [ 701.662550] env[61768]: DEBUG nova.compute.utils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.670213] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.670389] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 701.675940] env[61768]: DEBUG oslo_concurrency.lockutils [req-46fc32b3-7b21-4cd3-a100-a8b809038f84 req-e2aacf73-50fe-4775-9cb4-8ce8faa420a6 service nova] Releasing lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.676849] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Acquired lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.677053] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 701.718643] env[61768]: DEBUG nova.policy [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a4d3edcc6b5042a8bd7bc7cce597e8e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8700860fb1eb46e0afdd01ddf74b4e0c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.906394] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.039371] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d445ff7-7874-474a-8cc2-788f9d61dc47 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.050024] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80107902-d616-4e0b-8829-195fa64c446f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.089661] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Successfully created port: c4c98ad0-6b0f-4a0e-af7f-581439418979 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.092310] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7e7440-add2-4170-a3ca-05177ac3fbbe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.104019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b03a900c-1578-446a-94a2-ec71749694ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.116587] env[61768]: DEBUG nova.compute.provider_tree [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.170933] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.203350] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 702.304561] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.623111] env[61768]: DEBUG nova.scheduler.client.report [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.714113] env[61768]: DEBUG nova.compute.manager [req-389cccb3-46b0-40dc-a1e3-c992247a91a1 req-2d80bec9-5e23-437c-9670-1213c9696a4c service nova] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Received event network-vif-deleted-729d32df-bf2d-4fc2-ae66-57fca1610887 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.807183] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Releasing lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.807717] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 702.807959] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 702.808348] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c97061b-cb5a-4670-aa1e-e435eeddb5a8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.824382] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9a9acc3-619f-4238-bacd-ccc97376037e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.851633] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a83f5b0f-508c-4829-a4a6-641baa7b4c95 could not be found. [ 702.853443] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 702.853443] env[61768]: INFO nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Took 0.04 seconds to destroy the instance on the hypervisor. [ 702.853443] env[61768]: DEBUG oslo.service.loopingcall [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 702.853443] env[61768]: DEBUG nova.compute.manager [-] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 702.853443] env[61768]: DEBUG nova.network.neutron [-] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 702.871323] env[61768]: DEBUG nova.network.neutron [-] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.128538] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.129222] env[61768]: ERROR nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Traceback (most recent call last): [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self.driver.spawn(context, instance, image_meta, [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] vm_ref = self.build_virtual_machine(instance, [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.129222] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] for vif in network_info: [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] return self._sync_wrapper(fn, *args, **kwargs) [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self.wait() [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self[:] = self._gt.wait() [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] return self._exit_event.wait() [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] current.throw(*self._exc) [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.129578] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] result = function(*args, **kwargs) [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] return func(*args, **kwargs) [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] raise e [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] nwinfo = self.network_api.allocate_for_instance( [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] created_port_ids = self._update_ports_for_instance( [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] with excutils.save_and_reraise_exception(): [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] self.force_reraise() [ 703.129990] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] raise self.value [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] updated_port = self._update_port( [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] _ensure_no_port_binding_failure(port) [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] raise exception.PortBindingFailed(port_id=port['id']) [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] nova.exception.PortBindingFailed: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. [ 703.130390] env[61768]: ERROR nova.compute.manager [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] [ 703.130390] env[61768]: DEBUG nova.compute.utils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.131149] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.160s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.134601] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Build of instance 5da49801-148d-4b26-a0a2-816b1d5f7994 was re-scheduled: Binding failed for port 33194713-4bfb-4de3-9448-ed567fb0bde9, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 703.135059] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 703.135287] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquiring lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.135433] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Acquired lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.135591] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.157879] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.158120] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.185838] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.217330] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.217579] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.217738] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.217916] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.218073] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.218931] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.218931] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.218931] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.218931] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.218931] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.219177] env[61768]: DEBUG nova.virt.hardware [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.220341] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c56d69-4016-413b-9b65-947c8fb02f88 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.229268] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efb28a5-52e4-4bb6-b2c6-12328ab7bf7e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.369972] env[61768]: ERROR nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. [ 703.369972] env[61768]: ERROR nova.compute.manager Traceback (most recent call last): [ 703.369972] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.369972] env[61768]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 703.369972] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.369972] env[61768]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 703.369972] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.369972] env[61768]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 703.369972] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.369972] env[61768]: ERROR nova.compute.manager self.force_reraise() [ 703.369972] env[61768]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.369972] env[61768]: ERROR nova.compute.manager raise self.value [ 703.369972] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 703.369972] env[61768]: ERROR nova.compute.manager updated_port = self._update_port( [ 703.369972] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.369972] env[61768]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 703.370549] env[61768]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.370549] env[61768]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 703.370549] env[61768]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. [ 703.370549] env[61768]: ERROR nova.compute.manager [ 703.370549] env[61768]: Traceback (most recent call last): [ 703.370549] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 703.370549] env[61768]: listener.cb(fileno) [ 703.370549] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.370549] env[61768]: result = function(*args, **kwargs) [ 703.370549] env[61768]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.370549] env[61768]: return func(*args, **kwargs) [ 703.370549] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.370549] env[61768]: raise e [ 703.370549] env[61768]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.370549] env[61768]: nwinfo = self.network_api.allocate_for_instance( [ 703.370549] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.370549] env[61768]: created_port_ids = self._update_ports_for_instance( [ 703.370549] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.370549] env[61768]: with excutils.save_and_reraise_exception(): [ 703.370549] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.370549] env[61768]: self.force_reraise() [ 703.370549] env[61768]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.370549] env[61768]: raise self.value [ 703.370549] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 703.370549] env[61768]: updated_port = self._update_port( [ 703.370549] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.370549] env[61768]: _ensure_no_port_binding_failure(port) [ 703.370549] env[61768]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.370549] env[61768]: raise exception.PortBindingFailed(port_id=port['id']) [ 703.371451] env[61768]: nova.exception.PortBindingFailed: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. [ 703.371451] env[61768]: Removing descriptor: 17 [ 703.371451] env[61768]: ERROR nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Traceback (most recent call last): [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] yield resources [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self.driver.spawn(context, instance, image_meta, [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.371451] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] vm_ref = self.build_virtual_machine(instance, [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] for vif in network_info: [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return self._sync_wrapper(fn, *args, **kwargs) [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self.wait() [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self[:] = self._gt.wait() [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return self._exit_event.wait() [ 703.371827] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] result = hub.switch() [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return self.greenlet.switch() [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] result = function(*args, **kwargs) [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return func(*args, **kwargs) [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] raise e [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] nwinfo = self.network_api.allocate_for_instance( [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 703.372195] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] created_port_ids = self._update_ports_for_instance( [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] with excutils.save_and_reraise_exception(): [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self.force_reraise() [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] raise self.value [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] updated_port = self._update_port( [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] _ensure_no_port_binding_failure(port) [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.372578] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] raise exception.PortBindingFailed(port_id=port['id']) [ 703.372918] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] nova.exception.PortBindingFailed: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. [ 703.372918] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] [ 703.372918] env[61768]: INFO nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Terminating instance [ 703.373853] env[61768]: DEBUG nova.network.neutron [-] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.375094] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Acquiring lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.375255] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Acquired lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.375419] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.659388] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.755920] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.876981] env[61768]: INFO nova.compute.manager [-] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Took 1.02 seconds to deallocate network for instance. [ 703.882012] env[61768]: DEBUG nova.compute.claims [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 703.882213] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.894746] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.981307] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.002038] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40153e1c-cbcb-4b9d-8917-06ac070de504 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.010174] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95ee303d-3740-4d6a-8bd2-8146e4dbc4f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.041109] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129961a8-8c3b-4496-98ae-2533e08136e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.048970] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcfb2124-4abd-4abf-b2e1-d24ba0b436ee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.062328] env[61768]: DEBUG nova.compute.provider_tree [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.261602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Releasing lock "refresh_cache-5da49801-148d-4b26-a0a2-816b1d5f7994" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.261602] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 704.261602] env[61768]: DEBUG nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.261602] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.275870] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.485025] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Releasing lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.485025] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 704.485025] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 704.486027] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-64febe89-b2a0-4e6e-84ed-d91b926d0433 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.495065] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b85469d1-219d-4b42-a4d0-d92f20787110 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.515889] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f6b07607-632d-46d9-a72f-5a524fd20a3b could not be found. [ 704.516126] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 704.516342] env[61768]: INFO nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 704.516647] env[61768]: DEBUG oslo.service.loopingcall [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.516876] env[61768]: DEBUG nova.compute.manager [-] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.516967] env[61768]: DEBUG nova.network.neutron [-] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.531731] env[61768]: DEBUG nova.network.neutron [-] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.564336] env[61768]: DEBUG nova.scheduler.client.report [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.740757] env[61768]: DEBUG nova.compute.manager [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Received event network-changed-c4c98ad0-6b0f-4a0e-af7f-581439418979 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.740994] env[61768]: DEBUG nova.compute.manager [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Refreshing instance network info cache due to event network-changed-c4c98ad0-6b0f-4a0e-af7f-581439418979. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 704.741176] env[61768]: DEBUG oslo_concurrency.lockutils [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] Acquiring lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.741590] env[61768]: DEBUG oslo_concurrency.lockutils [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] Acquired lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.741807] env[61768]: DEBUG nova.network.neutron [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Refreshing network info cache for port c4c98ad0-6b0f-4a0e-af7f-581439418979 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 704.778955] env[61768]: DEBUG nova.network.neutron [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.034631] env[61768]: DEBUG nova.network.neutron [-] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.069596] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.069783] env[61768]: ERROR nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Traceback (most recent call last): [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self.driver.spawn(context, instance, image_meta, [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] vm_ref = self.build_virtual_machine(instance, [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] vif_infos = vmwarevif.get_vif_info(self._session, [ 705.069783] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] for vif in network_info: [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return self._sync_wrapper(fn, *args, **kwargs) [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self.wait() [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self[:] = self._gt.wait() [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return self._exit_event.wait() [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] result = hub.switch() [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 705.070193] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return self.greenlet.switch() [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] result = function(*args, **kwargs) [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] return func(*args, **kwargs) [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] raise e [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] nwinfo = self.network_api.allocate_for_instance( [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] created_port_ids = self._update_ports_for_instance( [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] with excutils.save_and_reraise_exception(): [ 705.070571] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] self.force_reraise() [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] raise self.value [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] updated_port = self._update_port( [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] _ensure_no_port_binding_failure(port) [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] raise exception.PortBindingFailed(port_id=port['id']) [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] nova.exception.PortBindingFailed: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. [ 705.070931] env[61768]: ERROR nova.compute.manager [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] [ 705.071259] env[61768]: DEBUG nova.compute.utils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 705.071653] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.384s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.073159] env[61768]: INFO nova.compute.claims [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.075946] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Build of instance c299abc5-7913-4eea-ac0a-02e713c6009c was re-scheduled: Binding failed for port 7bb429c1-2068-427b-9bd5-986393fcbb09, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 705.076417] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 705.076661] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Acquiring lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.076812] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Acquired lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.076971] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.260799] env[61768]: DEBUG nova.network.neutron [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.281281] env[61768]: INFO nova.compute.manager [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] [instance: 5da49801-148d-4b26-a0a2-816b1d5f7994] Took 1.02 seconds to deallocate network for instance. [ 705.332897] env[61768]: DEBUG nova.network.neutron [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.538224] env[61768]: INFO nova.compute.manager [-] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Took 1.02 seconds to deallocate network for instance. [ 705.540688] env[61768]: DEBUG nova.compute.claims [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Aborting claim: {{(pid=61768) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 705.540901] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.593380] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.658924] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.835719] env[61768]: DEBUG oslo_concurrency.lockutils [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] Releasing lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.835719] env[61768]: DEBUG nova.compute.manager [req-819841fd-a324-4a28-ba7b-411c3a92b343 req-0cce7857-1522-4bd5-b8a0-13997d67d629 service nova] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Received event network-vif-deleted-c4c98ad0-6b0f-4a0e-af7f-581439418979 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 706.160831] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Releasing lock "refresh_cache-c299abc5-7913-4eea-ac0a-02e713c6009c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.161827] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 706.161827] env[61768]: DEBUG nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 706.161827] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 706.178372] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 706.307036] env[61768]: INFO nova.scheduler.client.report [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Deleted allocations for instance 5da49801-148d-4b26-a0a2-816b1d5f7994 [ 706.365020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d83c849d-f094-4c2e-a09f-db875b06055f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.373708] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02fc381-0155-41f6-a394-1d953199f81b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.408711] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 706.409185] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 706.411011] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01414ef-adc3-472a-826d-cb1245ccf6bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.418662] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42bea1e-14c1-4b9d-8214-b06510978067 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.435794] env[61768]: DEBUG nova.compute.provider_tree [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.681153] env[61768]: DEBUG nova.network.neutron [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.816756] env[61768]: DEBUG oslo_concurrency.lockutils [None req-62d14209-21d7-4ed7-bd7a-9d73560bb33f tempest-VolumesAdminNegativeTest-1041300997 tempest-VolumesAdminNegativeTest-1041300997-project-member] Lock "5da49801-148d-4b26-a0a2-816b1d5f7994" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 155.542s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.918771] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 706.918771] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 706.918771] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Rebuilding the list of instances to heal {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 706.940927] env[61768]: DEBUG nova.scheduler.client.report [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.184028] env[61768]: INFO nova.compute.manager [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] [instance: c299abc5-7913-4eea-ac0a-02e713c6009c] Took 1.02 seconds to deallocate network for instance. [ 707.321184] env[61768]: DEBUG nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.421703] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.421876] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.422411] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 707.439987] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "refresh_cache-f81aabf6-ae46-405c-b101-12ca707a0567" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.440240] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquired lock "refresh_cache-f81aabf6-ae46-405c-b101-12ca707a0567" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.440322] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Forcefully refreshing network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 707.440416] env[61768]: DEBUG nova.objects.instance [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lazy-loading 'info_cache' on Instance uuid f81aabf6-ae46-405c-b101-12ca707a0567 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 707.450025] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.450211] env[61768]: DEBUG nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 707.452739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.042s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.454149] env[61768]: INFO nova.compute.claims [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.844573] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.958138] env[61768]: DEBUG nova.compute.utils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 707.961493] env[61768]: DEBUG nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 707.961687] env[61768]: DEBUG nova.network.neutron [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 708.047885] env[61768]: DEBUG nova.policy [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.229020] env[61768]: INFO nova.scheduler.client.report [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Deleted allocations for instance c299abc5-7913-4eea-ac0a-02e713c6009c [ 708.468762] env[61768]: DEBUG nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.470551] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.595310] env[61768]: DEBUG nova.network.neutron [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Successfully created port: 3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.738432] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6efbb35-9f95-4c17-9a44-907e80d5172f tempest-ServerActionsTestJSON-152570915 tempest-ServerActionsTestJSON-152570915-project-member] Lock "c299abc5-7913-4eea-ac0a-02e713c6009c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.459s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.802128] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d879336d-8ff2-43c2-93c0-ae66854cce81 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.811783] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f933f2-ba30-4610-93d9-083a8752e044 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.843039] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1f3f49-ce2d-4f01-a995-b2386ac2cd3a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.855540] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dabdcf9-3b40-4170-aa63-c47760a94ac4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.869610] env[61768]: DEBUG nova.compute.provider_tree [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.185670] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.241638] env[61768]: DEBUG nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 709.373095] env[61768]: DEBUG nova.scheduler.client.report [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.484784] env[61768]: DEBUG nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.524644] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.524859] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.525028] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.525217] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.525363] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.525550] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.525702] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.525860] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.526032] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.526389] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.526389] env[61768]: DEBUG nova.virt.hardware [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.527576] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c10f23-1713-496f-9a36-5f0df5a1f1c6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.536236] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d79c9c3-e4b3-451c-9557-27376e1ef31d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.689238] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Releasing lock "refresh_cache-f81aabf6-ae46-405c-b101-12ca707a0567" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.689424] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Updated the network info_cache for instance {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 709.689682] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.689910] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.690130] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.690344] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.690544] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.690803] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.691077] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 709.691299] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 709.769531] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.882036] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.882036] env[61768]: DEBUG nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 709.882036] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.585s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.884707] env[61768]: INFO nova.compute.claims [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 710.194615] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.391174] env[61768]: DEBUG nova.compute.utils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.394497] env[61768]: DEBUG nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.396722] env[61768]: DEBUG nova.network.neutron [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 710.475167] env[61768]: DEBUG nova.policy [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7f7389d526c4ce9bc8ecf062b9921e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '180da0bcd5954eb2bc89a20cf8f42bb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.898039] env[61768]: DEBUG nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 710.975097] env[61768]: DEBUG nova.compute.manager [req-2cb330fa-186d-47c9-a9d5-6a79f75a27d4 req-bb556b52-2d17-4797-86c7-ecb17846c63b service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-vif-plugged-3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 710.975097] env[61768]: DEBUG oslo_concurrency.lockutils [req-2cb330fa-186d-47c9-a9d5-6a79f75a27d4 req-bb556b52-2d17-4797-86c7-ecb17846c63b service nova] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.975097] env[61768]: DEBUG oslo_concurrency.lockutils [req-2cb330fa-186d-47c9-a9d5-6a79f75a27d4 req-bb556b52-2d17-4797-86c7-ecb17846c63b service nova] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.975097] env[61768]: DEBUG oslo_concurrency.lockutils [req-2cb330fa-186d-47c9-a9d5-6a79f75a27d4 req-bb556b52-2d17-4797-86c7-ecb17846c63b service nova] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.975097] env[61768]: DEBUG nova.compute.manager [req-2cb330fa-186d-47c9-a9d5-6a79f75a27d4 req-bb556b52-2d17-4797-86c7-ecb17846c63b service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] No waiting events found dispatching network-vif-plugged-3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 710.975515] env[61768]: WARNING nova.compute.manager [req-2cb330fa-186d-47c9-a9d5-6a79f75a27d4 req-bb556b52-2d17-4797-86c7-ecb17846c63b service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received unexpected event network-vif-plugged-3a76322f-8de0-44f9-9b42-ae3f6c592d3f for instance with vm_state building and task_state spawning. [ 711.201873] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15265c78-098e-481b-a945-d8cf5eed518b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.210262] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0918e7ab-6112-43c8-a415-0c2947f5c01e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.242246] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87ddeea8-bd14-48db-9b3e-af7aec81e17e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.250165] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b95c7ac-f481-4c2b-8d44-aae3301b334e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.263233] env[61768]: DEBUG nova.compute.provider_tree [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.302740] env[61768]: DEBUG nova.network.neutron [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Successfully updated port: 3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 711.312190] env[61768]: DEBUG nova.network.neutron [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Successfully created port: 59dd0ec8-7838-4fa2-ac70-42482a388241 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 711.770357] env[61768]: DEBUG nova.scheduler.client.report [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.807117] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.807117] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.807311] env[61768]: DEBUG nova.network.neutron [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.910712] env[61768]: DEBUG nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 711.942423] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 711.942668] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 711.942827] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 711.943119] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 711.943182] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 711.943320] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 711.943521] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 711.943678] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 711.943836] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 711.944010] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 711.947122] env[61768]: DEBUG nova.virt.hardware [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 711.948154] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b407c7-de1c-40ac-9653-49c3c9d0fc46 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.956260] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7c755b-62bd-421b-b52c-333ccb2641dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.278332] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.278825] env[61768]: DEBUG nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 712.281507] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.038s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.282915] env[61768]: INFO nova.compute.claims [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.344539] env[61768]: DEBUG nova.network.neutron [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.533245] env[61768]: DEBUG nova.network.neutron [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.787701] env[61768]: DEBUG nova.compute.utils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 712.791270] env[61768]: DEBUG nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 712.791452] env[61768]: DEBUG nova.network.neutron [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 712.868626] env[61768]: DEBUG nova.policy [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a2eb3cbf61bc4bdbaaaf29cdf78e4853', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76967cef64d8478ab599e0f336d67535', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 713.036580] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.036797] env[61768]: DEBUG nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Instance network_info: |[{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 713.038320] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:2f:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3a76322f-8de0-44f9-9b42-ae3f6c592d3f', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.045761] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Creating folder: Project (2f3e8b04445948a6a8a737ad89a5c002). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.046055] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee041eaa-0588-4d63-bc68-68ec27fb1b8e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.059990] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Created folder: Project (2f3e8b04445948a6a8a737ad89a5c002) in parent group-v265360. [ 713.059990] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Creating folder: Instances. Parent ref: group-v265378. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.059990] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d6ece70e-4172-4920-bd53-999c5ce269ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.069243] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Created folder: Instances in parent group-v265378. [ 713.069243] env[61768]: DEBUG oslo.service.loopingcall [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.069243] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 713.069243] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55474450-a0bd-4c2b-85cc-a8503e2f912e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.092434] env[61768]: DEBUG nova.compute.manager [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-changed-3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.092785] env[61768]: DEBUG nova.compute.manager [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing instance network info cache due to event network-changed-3a76322f-8de0-44f9-9b42-ae3f6c592d3f. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 713.093218] env[61768]: DEBUG oslo_concurrency.lockutils [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.093482] env[61768]: DEBUG oslo_concurrency.lockutils [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.093762] env[61768]: DEBUG nova.network.neutron [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing network info cache for port 3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.102127] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.102127] env[61768]: value = "task-1228608" [ 713.102127] env[61768]: _type = "Task" [ 713.102127] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.109909] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228608, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.296862] env[61768]: DEBUG nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 713.402159] env[61768]: DEBUG nova.network.neutron [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Successfully created port: 42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 713.610708] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228608, 'name': CreateVM_Task, 'duration_secs': 0.338836} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.613831] env[61768]: DEBUG nova.network.neutron [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Successfully updated port: 59dd0ec8-7838-4fa2-ac70-42482a388241 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 713.616693] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 713.630792] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.631022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.631532] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 713.631767] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17e03f84-23c8-4542-8e5f-54253cf88bfb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.637020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c2e1c3d-4252-46f4-919f-1e359cbbc4e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.642931] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 713.642931] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cd3590-68ca-a0ac-3919-af15cae254fa" [ 713.642931] env[61768]: _type = "Task" [ 713.642931] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.648705] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8c87548-e1e6-4b4b-b7fc-ff4141882a7b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.660172] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cd3590-68ca-a0ac-3919-af15cae254fa, 'name': SearchDatastore_Task, 'duration_secs': 0.011597} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.685712] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.686071] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 713.686392] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.686600] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.686897] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 713.687688] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ea654f77-35ed-475b-beb3-a3a18e54f87d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.690062] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1beffb-a1eb-4db1-9cc1-a28b43181df5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.697762] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2409c091-bdbf-4016-bf68-e85f205bd218 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.702541] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 713.702800] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 713.704120] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83d77041-1e6f-42c7-a235-d6826618ebb6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.714480] env[61768]: DEBUG nova.compute.provider_tree [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.719325] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 713.719325] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522c93b1-ef42-1d43-fcd7-148feb560136" [ 713.719325] env[61768]: _type = "Task" [ 713.719325] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.726028] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522c93b1-ef42-1d43-fcd7-148feb560136, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.119963] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "refresh_cache-663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.120138] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "refresh_cache-663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.120298] env[61768]: DEBUG nova.network.neutron [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 714.122274] env[61768]: DEBUG nova.network.neutron [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updated VIF entry in instance network info cache for port 3a76322f-8de0-44f9-9b42-ae3f6c592d3f. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 714.122586] env[61768]: DEBUG nova.network.neutron [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.217404] env[61768]: DEBUG nova.scheduler.client.report [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.236408] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522c93b1-ef42-1d43-fcd7-148feb560136, 'name': SearchDatastore_Task, 'duration_secs': 0.008917} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.238087] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a5d4c14-cfff-4410-ac9b-2f792b01e28f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.244232] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 714.244232] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5249120e-9190-43fb-6dbe-42a6e5d34a3f" [ 714.244232] env[61768]: _type = "Task" [ 714.244232] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.253495] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5249120e-9190-43fb-6dbe-42a6e5d34a3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.304105] env[61768]: DEBUG nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 714.340483] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 714.340733] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 714.340892] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 714.341083] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 714.344402] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 714.344935] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 714.344935] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 714.345131] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 714.345228] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 714.345421] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 714.345600] env[61768]: DEBUG nova.virt.hardware [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 714.346647] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af26b4aa-4631-47a2-9459-f1af9b27a695 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.355086] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efc59db-dbf7-4374-88d8-55c307679008 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.628235] env[61768]: DEBUG oslo_concurrency.lockutils [req-e93ca6a6-9400-487e-9a76-21202b7a525c req-04e54d5c-f8ca-4854-8d6b-29806b524a72 service nova] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.685284] env[61768]: DEBUG nova.network.neutron [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 714.730188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.730431] env[61768]: DEBUG nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.737019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.603s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.737019] env[61768]: INFO nova.compute.claims [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.755423] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5249120e-9190-43fb-6dbe-42a6e5d34a3f, 'name': SearchDatastore_Task, 'duration_secs': 0.010293} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.755688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.756308] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 12a543c2-1081-49d7-800b-07f0a2516904/12a543c2-1081-49d7-800b-07f0a2516904.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 714.756637] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5296180b-63eb-4b87-9424-5a0a8c032ddb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.767258] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 714.767258] env[61768]: value = "task-1228609" [ 714.767258] env[61768]: _type = "Task" [ 714.767258] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.778914] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228609, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.929618] env[61768]: DEBUG nova.network.neutron [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Updating instance_info_cache with network_info: [{"id": "59dd0ec8-7838-4fa2-ac70-42482a388241", "address": "fa:16:3e:92:ef:26", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59dd0ec8-78", "ovs_interfaceid": "59dd0ec8-7838-4fa2-ac70-42482a388241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.218265] env[61768]: DEBUG nova.compute.manager [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Received event network-vif-plugged-59dd0ec8-7838-4fa2-ac70-42482a388241 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.218476] env[61768]: DEBUG oslo_concurrency.lockutils [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] Acquiring lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.220016] env[61768]: DEBUG oslo_concurrency.lockutils [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.220016] env[61768]: DEBUG oslo_concurrency.lockutils [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.220016] env[61768]: DEBUG nova.compute.manager [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] No waiting events found dispatching network-vif-plugged-59dd0ec8-7838-4fa2-ac70-42482a388241 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 715.220016] env[61768]: WARNING nova.compute.manager [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Received unexpected event network-vif-plugged-59dd0ec8-7838-4fa2-ac70-42482a388241 for instance with vm_state building and task_state spawning. [ 715.220016] env[61768]: DEBUG nova.compute.manager [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Received event network-changed-59dd0ec8-7838-4fa2-ac70-42482a388241 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.220243] env[61768]: DEBUG nova.compute.manager [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Refreshing instance network info cache due to event network-changed-59dd0ec8-7838-4fa2-ac70-42482a388241. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 715.220243] env[61768]: DEBUG oslo_concurrency.lockutils [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] Acquiring lock "refresh_cache-663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.235638] env[61768]: DEBUG nova.compute.utils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.236966] env[61768]: DEBUG nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.237155] env[61768]: DEBUG nova.network.neutron [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 715.277385] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228609, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.462353} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.277844] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 12a543c2-1081-49d7-800b-07f0a2516904/12a543c2-1081-49d7-800b-07f0a2516904.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 715.278191] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 715.278454] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d06a29c-a9a1-4240-a02c-98fdfa4e4963 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.289463] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 715.289463] env[61768]: value = "task-1228610" [ 715.289463] env[61768]: _type = "Task" [ 715.289463] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.298181] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228610, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.313421] env[61768]: DEBUG nova.policy [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62522bd3731a47bc928a0c4038c447d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06ce9d5af7e740ff882439e4486a0aad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.435296] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "refresh_cache-663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.435635] env[61768]: DEBUG nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Instance network_info: |[{"id": "59dd0ec8-7838-4fa2-ac70-42482a388241", "address": "fa:16:3e:92:ef:26", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59dd0ec8-78", "ovs_interfaceid": "59dd0ec8-7838-4fa2-ac70-42482a388241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 715.435990] env[61768]: DEBUG oslo_concurrency.lockutils [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] Acquired lock "refresh_cache-663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.436297] env[61768]: DEBUG nova.network.neutron [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Refreshing network info cache for port 59dd0ec8-7838-4fa2-ac70-42482a388241 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 715.438038] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:ef:26', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59dd0ec8-7838-4fa2-ac70-42482a388241', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 715.451629] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating folder: Project (180da0bcd5954eb2bc89a20cf8f42bb5). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 715.454956] env[61768]: DEBUG nova.network.neutron [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Successfully updated port: 42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 715.458119] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-300f4aee-e838-473d-aab9-656699018c78 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.470860] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Created folder: Project (180da0bcd5954eb2bc89a20cf8f42bb5) in parent group-v265360. [ 715.471095] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating folder: Instances. Parent ref: group-v265381. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 715.471333] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5b8061dd-650b-4f96-8193-8363f5cbb33b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.480496] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Created folder: Instances in parent group-v265381. [ 715.484025] env[61768]: DEBUG oslo.service.loopingcall [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 715.484025] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 715.484025] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6c4d35b-cbdd-4e24-a7ec-e075ac504209 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.503060] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 715.503060] env[61768]: value = "task-1228613" [ 715.503060] env[61768]: _type = "Task" [ 715.503060] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.511249] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228613, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.741502] env[61768]: DEBUG nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.803562] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228610, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067969} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.810021] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 715.810021] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbf7e21-a319-44bd-93cb-2a8f53f7ada1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.836584] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 12a543c2-1081-49d7-800b-07f0a2516904/12a543c2-1081-49d7-800b-07f0a2516904.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 715.839801] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61ccea25-401c-4ba1-a644-ba544896ed2b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.860235] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 715.860235] env[61768]: value = "task-1228614" [ 715.860235] env[61768]: _type = "Task" [ 715.860235] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.869213] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228614, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.958627] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.958814] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquired lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.958969] env[61768]: DEBUG nova.network.neutron [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 716.020116] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228613, 'name': CreateVM_Task, 'duration_secs': 0.360398} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.022756] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 716.023568] env[61768]: DEBUG nova.network.neutron [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Successfully created port: 4b6df580-88b9-4455-b18b-0a48b05e7f73 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.026019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.026231] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.026564] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 716.027082] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfc1e371-0624-42e2-9946-90e433d67b6b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.031773] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 716.031773] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527df7ac-d9ca-25e4-83c6-000bf2659d8d" [ 716.031773] env[61768]: _type = "Task" [ 716.031773] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.043283] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527df7ac-d9ca-25e4-83c6-000bf2659d8d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.154347] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16e3e0bd-d6eb-41ef-9829-9c5c897fd9ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.161456] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d625256-f231-4aee-ab07-5f48927addc4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.198518] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2a6bf89-2bcb-4438-91d4-7fe3a9160c36 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.207444] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ce44d5-2dbe-4cd2-a095-498121073797 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.227452] env[61768]: DEBUG nova.compute.provider_tree [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.373167] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228614, 'name': ReconfigVM_Task, 'duration_secs': 0.288147} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.373504] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 12a543c2-1081-49d7-800b-07f0a2516904/12a543c2-1081-49d7-800b-07f0a2516904.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 716.374065] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9277a4e9-e435-499c-8cd3-52c8a1d2c839 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.381060] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 716.381060] env[61768]: value = "task-1228615" [ 716.381060] env[61768]: _type = "Task" [ 716.381060] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.386103] env[61768]: DEBUG nova.network.neutron [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Updated VIF entry in instance network info cache for port 59dd0ec8-7838-4fa2-ac70-42482a388241. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 716.386497] env[61768]: DEBUG nova.network.neutron [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Updating instance_info_cache with network_info: [{"id": "59dd0ec8-7838-4fa2-ac70-42482a388241", "address": "fa:16:3e:92:ef:26", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59dd0ec8-78", "ovs_interfaceid": "59dd0ec8-7838-4fa2-ac70-42482a388241", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.393222] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228615, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.542071] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527df7ac-d9ca-25e4-83c6-000bf2659d8d, 'name': SearchDatastore_Task, 'duration_secs': 0.00987} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.542475] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.542768] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 716.543030] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.543215] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.543392] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 716.543705] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4bfe17ab-f39f-4490-b3b8-e837436d181b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.570676] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 716.570858] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 716.571775] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-554a30d2-8997-4588-aeec-39219be7f61a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.575316] env[61768]: DEBUG nova.network.neutron [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.583041] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 716.583041] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52096385-e0f7-3652-7a00-7f26840eb5c2" [ 716.583041] env[61768]: _type = "Task" [ 716.583041] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.588484] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52096385-e0f7-3652-7a00-7f26840eb5c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.733069] env[61768]: DEBUG nova.scheduler.client.report [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.755682] env[61768]: DEBUG nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.789474] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.789634] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.789886] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.790165] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.790355] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.790733] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.790733] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.790922] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.791136] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.791339] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.791560] env[61768]: DEBUG nova.virt.hardware [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.792452] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ffe1531-5349-4ca4-aca7-c147d765aba1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.805217] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccf0b9e-836c-4ae3-ba02-15158253c4af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.830909] env[61768]: DEBUG nova.network.neutron [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updating instance_info_cache with network_info: [{"id": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "address": "fa:16:3e:c2:c0:a4", "network": {"id": "638c557a-bdca-4fa0-a3bb-ee6a1a5c4298", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1345341842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76967cef64d8478ab599e0f336d67535", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42b4fa2d-89", "ovs_interfaceid": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.894452] env[61768]: DEBUG oslo_concurrency.lockutils [req-29c85ae6-1981-4279-8f5a-9dce1e01cf4b req-72781c49-35fc-4a36-95b4-406ec4aaee5f service nova] Releasing lock "refresh_cache-663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.894866] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228615, 'name': Rename_Task, 'duration_secs': 0.175281} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.895293] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 716.895605] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8cb6860-a8d7-4960-9e6c-50ec5f363e92 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.903477] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 716.903477] env[61768]: value = "task-1228616" [ 716.903477] env[61768]: _type = "Task" [ 716.903477] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.912841] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.095128] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52096385-e0f7-3652-7a00-7f26840eb5c2, 'name': SearchDatastore_Task, 'duration_secs': 0.008259} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.098517] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfb98cb8-5c02-48a9-a762-2919bc5df800 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.106019] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 717.106019] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52dbb33b-18bf-ef50-1181-aab581c3b5f2" [ 717.106019] env[61768]: _type = "Task" [ 717.106019] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.116079] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52dbb33b-18bf-ef50-1181-aab581c3b5f2, 'name': SearchDatastore_Task, 'duration_secs': 0.00918} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.116394] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.116829] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5/663fa2fb-1ced-4052-9eba-8a5d0cbe96b5.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 717.117144] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c761c29-9f8d-406b-827a-c4765d30feaa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.127213] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 717.127213] env[61768]: value = "task-1228617" [ 717.127213] env[61768]: _type = "Task" [ 717.127213] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.136329] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.242035] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.242643] env[61768]: DEBUG nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.245842] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.844s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.247141] env[61768]: DEBUG nova.objects.instance [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 717.309385] env[61768]: DEBUG nova.compute.manager [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Received event network-vif-plugged-42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.309623] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] Acquiring lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.309835] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.310010] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.310189] env[61768]: DEBUG nova.compute.manager [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] No waiting events found dispatching network-vif-plugged-42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 717.310353] env[61768]: WARNING nova.compute.manager [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Received unexpected event network-vif-plugged-42b4fa2d-897f-4316-85f8-b2168ccfabdb for instance with vm_state building and task_state spawning. [ 717.310509] env[61768]: DEBUG nova.compute.manager [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Received event network-changed-42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.310660] env[61768]: DEBUG nova.compute.manager [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Refreshing instance network info cache due to event network-changed-42b4fa2d-897f-4316-85f8-b2168ccfabdb. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 717.310826] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] Acquiring lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.333192] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Releasing lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.333534] env[61768]: DEBUG nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Instance network_info: |[{"id": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "address": "fa:16:3e:c2:c0:a4", "network": {"id": "638c557a-bdca-4fa0-a3bb-ee6a1a5c4298", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1345341842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76967cef64d8478ab599e0f336d67535", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42b4fa2d-89", "ovs_interfaceid": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 717.333854] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] Acquired lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.334036] env[61768]: DEBUG nova.network.neutron [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Refreshing network info cache for port 42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 717.335345] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:c0:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '095fbf26-7367-4f9e-87c5-2965b64b0b0f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42b4fa2d-897f-4316-85f8-b2168ccfabdb', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.345131] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Creating folder: Project (76967cef64d8478ab599e0f336d67535). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.345667] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-351a6645-f342-4e07-8c30-495ac5882a66 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.359568] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Created folder: Project (76967cef64d8478ab599e0f336d67535) in parent group-v265360. [ 717.359568] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Creating folder: Instances. Parent ref: group-v265384. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.359892] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a219070a-f736-44f0-b271-de5847b1c1f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.371051] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Created folder: Instances in parent group-v265384. [ 717.371227] env[61768]: DEBUG oslo.service.loopingcall [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.371433] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 717.371640] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca36d2ec-b449-4c4b-9cef-bb35623d07fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.399126] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.399126] env[61768]: value = "task-1228620" [ 717.399126] env[61768]: _type = "Task" [ 717.399126] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.416057] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228620, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.421201] env[61768]: DEBUG oslo_vmware.api [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228616, 'name': PowerOnVM_Task, 'duration_secs': 0.443046} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.422134] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 717.422664] env[61768]: INFO nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Took 7.94 seconds to spawn the instance on the hypervisor. [ 717.422972] env[61768]: DEBUG nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 717.424119] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16031987-8077-413e-92af-b849aeaee450 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.639483] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.464891} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.639794] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5/663fa2fb-1ced-4052-9eba-8a5d0cbe96b5.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 717.640045] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.640323] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d71097cb-bf85-4b5d-827f-568d82349c12 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.651025] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 717.651025] env[61768]: value = "task-1228621" [ 717.651025] env[61768]: _type = "Task" [ 717.651025] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.662781] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228621, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.753635] env[61768]: DEBUG nova.compute.utils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.762021] env[61768]: DEBUG nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.762021] env[61768]: DEBUG nova.network.neutron [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 717.835669] env[61768]: DEBUG nova.policy [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '15d5b1eba3474b3a92b71fa8159c3504', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b5f4ceb636d143faaf0595e25cd6c01b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.917310] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228620, 'name': CreateVM_Task, 'duration_secs': 0.375135} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.919886] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 717.921174] env[61768]: DEBUG nova.compute.manager [req-ac0c016f-ee43-46ce-add7-fa903c60aa7c req-6847ad70-ab27-4a45-bc0e-f4e46dc0acac service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Received event network-vif-plugged-4b6df580-88b9-4455-b18b-0a48b05e7f73 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.921255] env[61768]: DEBUG oslo_concurrency.lockutils [req-ac0c016f-ee43-46ce-add7-fa903c60aa7c req-6847ad70-ab27-4a45-bc0e-f4e46dc0acac service nova] Acquiring lock "0df161e9-4a66-432e-9366-60f231c53e3d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.921494] env[61768]: DEBUG oslo_concurrency.lockutils [req-ac0c016f-ee43-46ce-add7-fa903c60aa7c req-6847ad70-ab27-4a45-bc0e-f4e46dc0acac service nova] Lock "0df161e9-4a66-432e-9366-60f231c53e3d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.921611] env[61768]: DEBUG oslo_concurrency.lockutils [req-ac0c016f-ee43-46ce-add7-fa903c60aa7c req-6847ad70-ab27-4a45-bc0e-f4e46dc0acac service nova] Lock "0df161e9-4a66-432e-9366-60f231c53e3d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.921781] env[61768]: DEBUG nova.compute.manager [req-ac0c016f-ee43-46ce-add7-fa903c60aa7c req-6847ad70-ab27-4a45-bc0e-f4e46dc0acac service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] No waiting events found dispatching network-vif-plugged-4b6df580-88b9-4455-b18b-0a48b05e7f73 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 717.921941] env[61768]: WARNING nova.compute.manager [req-ac0c016f-ee43-46ce-add7-fa903c60aa7c req-6847ad70-ab27-4a45-bc0e-f4e46dc0acac service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Received unexpected event network-vif-plugged-4b6df580-88b9-4455-b18b-0a48b05e7f73 for instance with vm_state building and task_state spawning. [ 717.925632] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.925632] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.925632] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 717.926361] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30b2e8c5-059e-4095-bb25-f1c829bd1c64 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.933080] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 717.933080] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52764d1c-d66d-983f-d3e4-09a6c14f8339" [ 717.933080] env[61768]: _type = "Task" [ 717.933080] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.944264] env[61768]: DEBUG nova.network.neutron [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Successfully updated port: 4b6df580-88b9-4455-b18b-0a48b05e7f73 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 717.960334] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52764d1c-d66d-983f-d3e4-09a6c14f8339, 'name': SearchDatastore_Task, 'duration_secs': 0.00995} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.962042] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.962042] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.962415] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.962603] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.962892] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.965945] env[61768]: INFO nova.compute.manager [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Took 28.29 seconds to build instance. [ 717.966917] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8cca272-9e5f-4938-bbb3-368c35a561ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.978809] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 717.979349] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 717.979768] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2b207cd-8b8d-4bd6-a604-aef6e3dd1676 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.988488] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 717.988488] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cd11f7-21cc-c730-62d5-f2c107d9b6c0" [ 717.988488] env[61768]: _type = "Task" [ 717.988488] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.998749] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cd11f7-21cc-c730-62d5-f2c107d9b6c0, 'name': SearchDatastore_Task, 'duration_secs': 0.009325} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.000322] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca481d70-aed2-44ea-afd9-747a41e1cee0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.005965] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 718.005965] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5222621d-dcca-0355-1100-cb19eb0ad161" [ 718.005965] env[61768]: _type = "Task" [ 718.005965] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.015374] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5222621d-dcca-0355-1100-cb19eb0ad161, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.162314] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228621, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071228} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.162314] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 718.162314] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e7a2f1-dffc-4458-9f4e-0613275a6409 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.186682] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5/663fa2fb-1ced-4052-9eba-8a5d0cbe96b5.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 718.187030] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-585bd05b-bcf5-480c-8821-802c09ecffbb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.215671] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 718.215671] env[61768]: value = "task-1228622" [ 718.215671] env[61768]: _type = "Task" [ 718.215671] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.227050] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228622, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.262121] env[61768]: DEBUG nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.264929] env[61768]: DEBUG oslo_concurrency.lockutils [None req-064dabf8-f2de-415e-847b-194d90b2814e tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.266037] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.079s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.266246] env[61768]: DEBUG nova.objects.instance [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lazy-loading 'resources' on Instance uuid f81aabf6-ae46-405c-b101-12ca707a0567 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 718.294054] env[61768]: DEBUG nova.network.neutron [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updated VIF entry in instance network info cache for port 42b4fa2d-897f-4316-85f8-b2168ccfabdb. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 718.294433] env[61768]: DEBUG nova.network.neutron [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updating instance_info_cache with network_info: [{"id": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "address": "fa:16:3e:c2:c0:a4", "network": {"id": "638c557a-bdca-4fa0-a3bb-ee6a1a5c4298", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1345341842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76967cef64d8478ab599e0f336d67535", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42b4fa2d-89", "ovs_interfaceid": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.423574] env[61768]: DEBUG nova.network.neutron [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Successfully created port: 90ce2022-9af9-4df8-b92c-140069264092 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.438212] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.438463] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.459770] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "refresh_cache-0df161e9-4a66-432e-9366-60f231c53e3d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.459770] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "refresh_cache-0df161e9-4a66-432e-9366-60f231c53e3d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 718.459770] env[61768]: DEBUG nova.network.neutron [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 718.471405] env[61768]: DEBUG oslo_concurrency.lockutils [None req-baeba34d-75fd-4815-bc45-d73270f9f330 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "12a543c2-1081-49d7-800b-07f0a2516904" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 149.708s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.519883] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5222621d-dcca-0355-1100-cb19eb0ad161, 'name': SearchDatastore_Task, 'duration_secs': 0.00799} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.519883] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.519883] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed/7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 718.519883] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6332b0bd-756d-43e5-8864-d80265cc0ba0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.524647] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 718.524647] env[61768]: value = "task-1228623" [ 718.524647] env[61768]: _type = "Task" [ 718.524647] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.532869] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228623, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.728831] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228622, 'name': ReconfigVM_Task, 'duration_secs': 0.268506} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.729203] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5/663fa2fb-1ced-4052-9eba-8a5d0cbe96b5.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.729949] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85fa2291-a577-4ab2-8a52-e451fdfddf3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.740321] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 718.740321] env[61768]: value = "task-1228624" [ 718.740321] env[61768]: _type = "Task" [ 718.740321] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.752141] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228624, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.798635] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c34256f-45ba-47aa-883f-6ea483169982 req-fa54a7b7-935f-417e-bf9d-6173776303b6 service nova] Releasing lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.975439] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 719.045137] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228623, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452292} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.046481] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed/7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 719.046726] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 719.046979] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0ec0d300-eee6-466c-ab76-b9086d48b2e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.054498] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 719.054498] env[61768]: value = "task-1228625" [ 719.054498] env[61768]: _type = "Task" [ 719.054498] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.066096] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228625, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.173566] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4d6acd-2294-4354-a9cd-4637f7f33807 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.181133] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fdf42df-16d8-4f50-a9e6-056051d208eb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.215881] env[61768]: DEBUG nova.network.neutron [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 719.217092] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54622276-5027-4a2a-a895-f3e6f9e10a95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.225127] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f57f80-d82f-433a-bd70-a16a86521989 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.239784] env[61768]: DEBUG nova.compute.provider_tree [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.257444] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228624, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.278988] env[61768]: DEBUG nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.299257] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.299580] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.310084] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.310338] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.310966] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.311314] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.311480] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.311707] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.311923] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.312096] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.312265] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.312426] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.312598] env[61768]: DEBUG nova.virt.hardware [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.313460] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcca5df4-5bb6-4b31-ba49-0df2ee539ddd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.322120] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca362a4-fa78-4319-9d0d-3323d1bc8c4c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.398099] env[61768]: DEBUG nova.network.neutron [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Updating instance_info_cache with network_info: [{"id": "4b6df580-88b9-4455-b18b-0a48b05e7f73", "address": "fa:16:3e:f8:c0:f8", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b6df580-88", "ovs_interfaceid": "4b6df580-88b9-4455-b18b-0a48b05e7f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.498981] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.564653] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228625, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066499} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.564937] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 719.565733] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2393693-2f47-4a6d-9e52-9fa13de19909 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.588553] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed/7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 719.589255] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2b9fd73-5152-4359-9dd9-0cf3124fb0fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.612451] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 719.612451] env[61768]: value = "task-1228626" [ 719.612451] env[61768]: _type = "Task" [ 719.612451] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.624140] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228626, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.748728] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228624, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.751737] env[61768]: DEBUG nova.scheduler.client.report [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.901616] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "refresh_cache-0df161e9-4a66-432e-9366-60f231c53e3d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.901960] env[61768]: DEBUG nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Instance network_info: |[{"id": "4b6df580-88b9-4455-b18b-0a48b05e7f73", "address": "fa:16:3e:f8:c0:f8", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b6df580-88", "ovs_interfaceid": "4b6df580-88b9-4455-b18b-0a48b05e7f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 719.903211] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:c0:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3291573-fad8-48cc-a965-c3554e7cee4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4b6df580-88b9-4455-b18b-0a48b05e7f73', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 719.913819] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Creating folder: Project (06ce9d5af7e740ff882439e4486a0aad). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 719.914598] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f763a3e-76f5-4f40-8299-4fc6340d0f70 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.926136] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Created folder: Project (06ce9d5af7e740ff882439e4486a0aad) in parent group-v265360. [ 719.928470] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Creating folder: Instances. Parent ref: group-v265387. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 719.928470] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a913ae85-9238-4484-955d-c4f4c89f1925 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.937023] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Created folder: Instances in parent group-v265387. [ 719.940017] env[61768]: DEBUG oslo.service.loopingcall [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 719.940017] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 719.940017] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0343bc40-d5fd-4995-9a50-4ff00f219516 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.959863] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 719.959863] env[61768]: value = "task-1228629" [ 719.959863] env[61768]: _type = "Task" [ 719.959863] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.968901] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228629, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.982765] env[61768]: DEBUG nova.compute.manager [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Received event network-changed-4b6df580-88b9-4455-b18b-0a48b05e7f73 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 719.983060] env[61768]: DEBUG nova.compute.manager [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Refreshing instance network info cache due to event network-changed-4b6df580-88b9-4455-b18b-0a48b05e7f73. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 719.983192] env[61768]: DEBUG oslo_concurrency.lockutils [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] Acquiring lock "refresh_cache-0df161e9-4a66-432e-9366-60f231c53e3d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.983347] env[61768]: DEBUG oslo_concurrency.lockutils [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] Acquired lock "refresh_cache-0df161e9-4a66-432e-9366-60f231c53e3d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.983563] env[61768]: DEBUG nova.network.neutron [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Refreshing network info cache for port 4b6df580-88b9-4455-b18b-0a48b05e7f73 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.094872] env[61768]: DEBUG nova.compute.manager [req-02d70122-f377-40c2-a810-58120bf0e362 req-b176589a-a21b-483a-87d7-fba2c91c7a0e service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Received event network-vif-plugged-90ce2022-9af9-4df8-b92c-140069264092 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.095171] env[61768]: DEBUG oslo_concurrency.lockutils [req-02d70122-f377-40c2-a810-58120bf0e362 req-b176589a-a21b-483a-87d7-fba2c91c7a0e service nova] Acquiring lock "3401be45-c3ab-4780-81a7-a93b05742414-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.095447] env[61768]: DEBUG oslo_concurrency.lockutils [req-02d70122-f377-40c2-a810-58120bf0e362 req-b176589a-a21b-483a-87d7-fba2c91c7a0e service nova] Lock "3401be45-c3ab-4780-81a7-a93b05742414-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.095665] env[61768]: DEBUG oslo_concurrency.lockutils [req-02d70122-f377-40c2-a810-58120bf0e362 req-b176589a-a21b-483a-87d7-fba2c91c7a0e service nova] Lock "3401be45-c3ab-4780-81a7-a93b05742414-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.095889] env[61768]: DEBUG nova.compute.manager [req-02d70122-f377-40c2-a810-58120bf0e362 req-b176589a-a21b-483a-87d7-fba2c91c7a0e service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] No waiting events found dispatching network-vif-plugged-90ce2022-9af9-4df8-b92c-140069264092 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 720.096135] env[61768]: WARNING nova.compute.manager [req-02d70122-f377-40c2-a810-58120bf0e362 req-b176589a-a21b-483a-87d7-fba2c91c7a0e service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Received unexpected event network-vif-plugged-90ce2022-9af9-4df8-b92c-140069264092 for instance with vm_state building and task_state spawning. [ 720.123569] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228626, 'name': ReconfigVM_Task, 'duration_secs': 0.276231} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.123961] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Reconfigured VM instance instance-0000002e to attach disk [datastore2] 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed/7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 720.125159] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e07aa780-b74e-47aa-9ec5-4a050aa62c4d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.132022] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 720.132022] env[61768]: value = "task-1228630" [ 720.132022] env[61768]: _type = "Task" [ 720.132022] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.146713] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228630, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.199688] env[61768]: DEBUG nova.network.neutron [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Successfully updated port: 90ce2022-9af9-4df8-b92c-140069264092 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 720.249600] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228624, 'name': Rename_Task, 'duration_secs': 1.329831} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.249889] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.250165] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-85ad376c-eec0-42df-b2a1-c5fb55970130 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.256217] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 720.256217] env[61768]: value = "task-1228631" [ 720.256217] env[61768]: _type = "Task" [ 720.256217] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.256982] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.991s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.262246] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.103s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.263862] env[61768]: INFO nova.compute.claims [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 720.271954] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.283680] env[61768]: INFO nova.scheduler.client.report [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Deleted allocations for instance f81aabf6-ae46-405c-b101-12ca707a0567 [ 720.470206] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228629, 'name': CreateVM_Task, 'duration_secs': 0.371451} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.470379] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 720.471141] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.471300] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.471607] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 720.471864] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeeb500d-12f7-43ab-af3b-766dd4f39bb0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.477571] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 720.477571] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5208335a-d4db-fabe-da6a-ade4b0a40ed2" [ 720.477571] env[61768]: _type = "Task" [ 720.477571] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.485896] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5208335a-d4db-fabe-da6a-ade4b0a40ed2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.642390] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228630, 'name': Rename_Task, 'duration_secs': 0.196472} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.642669] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 720.642729] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bafc685-37cf-43ca-9093-7e92b1dee94b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.648631] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 720.648631] env[61768]: value = "task-1228632" [ 720.648631] env[61768]: _type = "Task" [ 720.648631] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.659323] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228632, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.702014] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "refresh_cache-3401be45-c3ab-4780-81a7-a93b05742414" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.702188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquired lock "refresh_cache-3401be45-c3ab-4780-81a7-a93b05742414" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.702342] env[61768]: DEBUG nova.network.neutron [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 720.769323] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228631, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.778804] env[61768]: DEBUG nova.network.neutron [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Updated VIF entry in instance network info cache for port 4b6df580-88b9-4455-b18b-0a48b05e7f73. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 720.779213] env[61768]: DEBUG nova.network.neutron [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Updating instance_info_cache with network_info: [{"id": "4b6df580-88b9-4455-b18b-0a48b05e7f73", "address": "fa:16:3e:f8:c0:f8", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4b6df580-88", "ovs_interfaceid": "4b6df580-88b9-4455-b18b-0a48b05e7f73", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.792041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9e564f15-729d-429a-83d0-adda0c772458 tempest-ServerShowV247Test-1874716226 tempest-ServerShowV247Test-1874716226-project-member] Lock "f81aabf6-ae46-405c-b101-12ca707a0567" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.384s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.987893] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5208335a-d4db-fabe-da6a-ade4b0a40ed2, 'name': SearchDatastore_Task, 'duration_secs': 0.013019} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.988266] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.988533] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 720.988782] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.988928] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.989121] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 720.989379] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45629dbc-4669-449f-bb29-1da05a019bc0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.998649] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 720.998834] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 720.999569] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeafbed8-3a18-49b4-b657-b84e87137bad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.004813] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 721.004813] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5252914f-96a4-445f-de90-50d70697ae5d" [ 721.004813] env[61768]: _type = "Task" [ 721.004813] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.012703] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5252914f-96a4-445f-de90-50d70697ae5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.158850] env[61768]: DEBUG oslo_vmware.api [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228632, 'name': PowerOnVM_Task, 'duration_secs': 0.455597} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.159135] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.159344] env[61768]: INFO nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Took 6.86 seconds to spawn the instance on the hypervisor. [ 721.159521] env[61768]: DEBUG nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.160286] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0311f9e-af6b-4980-8844-9654400ea58c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.246533] env[61768]: DEBUG nova.network.neutron [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 721.266682] env[61768]: DEBUG oslo_vmware.api [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228631, 'name': PowerOnVM_Task, 'duration_secs': 0.550951} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.268809] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 721.269009] env[61768]: INFO nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Took 9.36 seconds to spawn the instance on the hypervisor. [ 721.269231] env[61768]: DEBUG nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.269978] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08544eeb-63e0-498d-88ee-8ce7ec445ab1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.283357] env[61768]: DEBUG oslo_concurrency.lockutils [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] Releasing lock "refresh_cache-0df161e9-4a66-432e-9366-60f231c53e3d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.283592] env[61768]: DEBUG nova.compute.manager [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-changed-3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.283761] env[61768]: DEBUG nova.compute.manager [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing instance network info cache due to event network-changed-3a76322f-8de0-44f9-9b42-ae3f6c592d3f. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 721.283968] env[61768]: DEBUG oslo_concurrency.lockutils [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.284123] env[61768]: DEBUG oslo_concurrency.lockutils [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.284283] env[61768]: DEBUG nova.network.neutron [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing network info cache for port 3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 721.421352] env[61768]: DEBUG nova.network.neutron [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Updating instance_info_cache with network_info: [{"id": "90ce2022-9af9-4df8-b92c-140069264092", "address": "fa:16:3e:e6:98:ad", "network": {"id": "0a890c79-4f71-416f-9f5f-6c0c0c66336d", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-355580779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5f4ceb636d143faaf0595e25cd6c01b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3117b312-701b-4439-b197-96b6c5cdca89", "external-id": "nsx-vlan-transportzone-94", "segmentation_id": 94, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ce2022-9a", "ovs_interfaceid": "90ce2022-9af9-4df8-b92c-140069264092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.517947] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5252914f-96a4-445f-de90-50d70697ae5d, 'name': SearchDatastore_Task, 'duration_secs': 0.039664} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.521063] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1898dc82-584c-4833-9c1f-4ca3ebf8b08a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.527560] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 721.527560] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e2d1de-2dcd-a40a-7328-f2dac912e3f8" [ 721.527560] env[61768]: _type = "Task" [ 721.527560] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.536016] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e2d1de-2dcd-a40a-7328-f2dac912e3f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.619409] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68423406-bd2a-4348-9c07-24fe952e8cfb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.627221] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-060ef79b-b27b-4a45-b912-0750b2261f35 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.658546] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0960c18-44e5-4101-b262-43bbf0f1822a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.666604] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08252d7e-08ec-4ecb-a7cd-16339dc34b49 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.687492] env[61768]: DEBUG nova.compute.provider_tree [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.688838] env[61768]: INFO nova.compute.manager [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Took 29.42 seconds to build instance. [ 721.791051] env[61768]: INFO nova.compute.manager [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Took 31.40 seconds to build instance. [ 721.924788] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Releasing lock "refresh_cache-3401be45-c3ab-4780-81a7-a93b05742414" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.925218] env[61768]: DEBUG nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Instance network_info: |[{"id": "90ce2022-9af9-4df8-b92c-140069264092", "address": "fa:16:3e:e6:98:ad", "network": {"id": "0a890c79-4f71-416f-9f5f-6c0c0c66336d", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-355580779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5f4ceb636d143faaf0595e25cd6c01b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3117b312-701b-4439-b197-96b6c5cdca89", "external-id": "nsx-vlan-transportzone-94", "segmentation_id": 94, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ce2022-9a", "ovs_interfaceid": "90ce2022-9af9-4df8-b92c-140069264092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 721.925689] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:98:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3117b312-701b-4439-b197-96b6c5cdca89', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '90ce2022-9af9-4df8-b92c-140069264092', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 721.935220] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Creating folder: Project (b5f4ceb636d143faaf0595e25cd6c01b). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.936045] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3f5d0a1-659d-4c66-af23-a22ab686b842 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.948561] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Created folder: Project (b5f4ceb636d143faaf0595e25cd6c01b) in parent group-v265360. [ 721.948561] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Creating folder: Instances. Parent ref: group-v265390. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 721.948757] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1c1be17c-06b2-4c18-9662-6e6d9fb13223 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.957886] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Created folder: Instances in parent group-v265390. [ 721.958302] env[61768]: DEBUG oslo.service.loopingcall [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 721.958510] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 721.958712] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-05b8fbb8-c7df-49a7-8c58-f362d995aa64 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.981293] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 721.981293] env[61768]: value = "task-1228635" [ 721.981293] env[61768]: _type = "Task" [ 721.981293] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.991130] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228635, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.039720] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e2d1de-2dcd-a40a-7328-f2dac912e3f8, 'name': SearchDatastore_Task, 'duration_secs': 0.030508} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.039986] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.040271] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0df161e9-4a66-432e-9366-60f231c53e3d/0df161e9-4a66-432e-9366-60f231c53e3d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 722.040528] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-480d6e46-37d6-4060-9a79-a164caa413d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.047861] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 722.047861] env[61768]: value = "task-1228636" [ 722.047861] env[61768]: _type = "Task" [ 722.047861] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.058763] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228636, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.122368] env[61768]: DEBUG nova.network.neutron [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updated VIF entry in instance network info cache for port 3a76322f-8de0-44f9-9b42-ae3f6c592d3f. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 722.122797] env[61768]: DEBUG nova.network.neutron [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.190772] env[61768]: DEBUG nova.scheduler.client.report [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 722.194565] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dc674a00-4d60-444c-9386-2ef8efa2099b tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.279s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.232923] env[61768]: DEBUG nova.compute.manager [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Received event network-changed-90ce2022-9af9-4df8-b92c-140069264092 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.233199] env[61768]: DEBUG nova.compute.manager [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Refreshing instance network info cache due to event network-changed-90ce2022-9af9-4df8-b92c-140069264092. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.233422] env[61768]: DEBUG oslo_concurrency.lockutils [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] Acquiring lock "refresh_cache-3401be45-c3ab-4780-81a7-a93b05742414" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.233596] env[61768]: DEBUG oslo_concurrency.lockutils [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] Acquired lock "refresh_cache-3401be45-c3ab-4780-81a7-a93b05742414" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.233975] env[61768]: DEBUG nova.network.neutron [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Refreshing network info cache for port 90ce2022-9af9-4df8-b92c-140069264092 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 722.247211] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7337b857-39da-4e75-b796-6d6ba6df6422 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.254385] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3b656738-f62f-4f3c-8c0d-21a70b72e2de tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Suspending the VM {{(pid=61768) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 722.255105] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-aa0d42ca-afe5-4350-9f5c-1fc04cccf09a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.263988] env[61768]: DEBUG oslo_vmware.api [None req-3b656738-f62f-4f3c-8c0d-21a70b72e2de tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 722.263988] env[61768]: value = "task-1228637" [ 722.263988] env[61768]: _type = "Task" [ 722.263988] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.275316] env[61768]: DEBUG oslo_vmware.api [None req-3b656738-f62f-4f3c-8c0d-21a70b72e2de tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228637, 'name': SuspendVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.293562] env[61768]: DEBUG oslo_concurrency.lockutils [None req-27b628a9-30b2-442d-8ca0-29faf3537841 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.619s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.375731] env[61768]: DEBUG nova.compute.manager [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Received event network-changed-42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 722.375731] env[61768]: DEBUG nova.compute.manager [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Refreshing instance network info cache due to event network-changed-42b4fa2d-897f-4316-85f8-b2168ccfabdb. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 722.376248] env[61768]: DEBUG oslo_concurrency.lockutils [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] Acquiring lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.376248] env[61768]: DEBUG oslo_concurrency.lockutils [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] Acquired lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.376325] env[61768]: DEBUG nova.network.neutron [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Refreshing network info cache for port 42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 722.492442] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228635, 'name': CreateVM_Task, 'duration_secs': 0.391519} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.492695] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 722.493584] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.493584] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.493838] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 722.494127] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f172ba5a-4900-4cb6-9bbc-0d5297163354 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.499662] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 722.499662] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5241fedd-166e-332f-9dad-70b65ababa0e" [ 722.499662] env[61768]: _type = "Task" [ 722.499662] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.507635] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5241fedd-166e-332f-9dad-70b65ababa0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.560135] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228636, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.625614] env[61768]: DEBUG oslo_concurrency.lockutils [req-06b6aaa9-1b2d-4184-a3b4-81010fd17c69 req-42bd81b3-4891-4bd4-b4fa-e6d2c2f4da45 service nova] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.698709] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.436s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.699855] env[61768]: DEBUG nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 722.703341] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 722.706152] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.800s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.706432] env[61768]: DEBUG nova.objects.instance [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lazy-loading 'resources' on Instance uuid 63704075-5eaf-4f84-a90b-3a0a3e904a9d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 722.775402] env[61768]: DEBUG oslo_vmware.api [None req-3b656738-f62f-4f3c-8c0d-21a70b72e2de tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228637, 'name': SuspendVM_Task} progress is 45%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.799024] env[61768]: DEBUG nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.011463] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5241fedd-166e-332f-9dad-70b65ababa0e, 'name': SearchDatastore_Task, 'duration_secs': 0.040582} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.011828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.012116] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 723.012397] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.012582] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.012800] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 723.013106] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-44edfa98-0d18-4761-9f1c-7914a98dd91a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.024716] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 723.024935] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 723.025766] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.025989] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.026226] env[61768]: INFO nova.compute.manager [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Rebooting instance [ 723.027614] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cec6416e-9c0c-4d58-adce-9d30acfff658 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.035540] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 723.035540] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528dd8b1-78f9-5bf5-a2c8-5e19af009ddb" [ 723.035540] env[61768]: _type = "Task" [ 723.035540] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.044235] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528dd8b1-78f9-5bf5-a2c8-5e19af009ddb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.058467] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228636, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.805212} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.058723] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0df161e9-4a66-432e-9366-60f231c53e3d/0df161e9-4a66-432e-9366-60f231c53e3d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 723.059190] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 723.059190] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e4190fb-7f7e-42b3-9dc7-a84299abb1e7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.066622] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 723.066622] env[61768]: value = "task-1228638" [ 723.066622] env[61768]: _type = "Task" [ 723.066622] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.075576] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228638, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.171481] env[61768]: DEBUG nova.network.neutron [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Updated VIF entry in instance network info cache for port 90ce2022-9af9-4df8-b92c-140069264092. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 723.171835] env[61768]: DEBUG nova.network.neutron [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Updating instance_info_cache with network_info: [{"id": "90ce2022-9af9-4df8-b92c-140069264092", "address": "fa:16:3e:e6:98:ad", "network": {"id": "0a890c79-4f71-416f-9f5f-6c0c0c66336d", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-355580779-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b5f4ceb636d143faaf0595e25cd6c01b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3117b312-701b-4439-b197-96b6c5cdca89", "external-id": "nsx-vlan-transportzone-94", "segmentation_id": 94, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap90ce2022-9a", "ovs_interfaceid": "90ce2022-9af9-4df8-b92c-140069264092", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.207850] env[61768]: DEBUG nova.compute.utils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 723.209387] env[61768]: DEBUG nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 723.210189] env[61768]: DEBUG nova.network.neutron [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 723.241365] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.276386] env[61768]: DEBUG oslo_vmware.api [None req-3b656738-f62f-4f3c-8c0d-21a70b72e2de tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228637, 'name': SuspendVM_Task, 'duration_secs': 0.937111} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.276654] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3b656738-f62f-4f3c-8c0d-21a70b72e2de tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Suspended the VM {{(pid=61768) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 723.276868] env[61768]: DEBUG nova.compute.manager [None req-3b656738-f62f-4f3c-8c0d-21a70b72e2de tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 723.277646] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff6e924-cf60-4964-a125-10f2356ffd9b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.295057] env[61768]: DEBUG nova.network.neutron [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updated VIF entry in instance network info cache for port 42b4fa2d-897f-4316-85f8-b2168ccfabdb. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 723.295448] env[61768]: DEBUG nova.network.neutron [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updating instance_info_cache with network_info: [{"id": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "address": "fa:16:3e:c2:c0:a4", "network": {"id": "638c557a-bdca-4fa0-a3bb-ee6a1a5c4298", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1345341842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76967cef64d8478ab599e0f336d67535", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42b4fa2d-89", "ovs_interfaceid": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.300787] env[61768]: DEBUG nova.policy [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f41abef76a5b43c2a222b952d8a8d05a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55c519433a0e4f2b85fec7f8ea820496', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 723.321412] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.513256] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d2106e7-fff5-4591-9da1-76c12e25e46b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.521466] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c821b59-5c3a-4b9d-a3ef-b244de2ebfa4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.571741] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ab6641-433e-4257-a002-b164756ad2de {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.585137] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528dd8b1-78f9-5bf5-a2c8-5e19af009ddb, 'name': SearchDatastore_Task, 'duration_secs': 0.010358} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.588122] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-220af862-2df6-4157-a838-4aaf1a5ddf51 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.591743] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5318050d-015f-4c8b-a5fb-83356d43e257 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.600094] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.601028] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228638, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099124} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.610548] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 723.610956] env[61768]: DEBUG nova.compute.provider_tree [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.612247] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 723.612247] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526c704c-8db0-ff1e-91f3-b0a026121995" [ 723.612247] env[61768]: _type = "Task" [ 723.612247] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.616451] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d974e95-9e66-47ee-8cfb-ca87fc8513e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.616948] env[61768]: DEBUG nova.network.neutron [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Successfully created port: 932b7671-16df-463c-b31c-e99ede6a0287 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.627853] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526c704c-8db0-ff1e-91f3-b0a026121995, 'name': SearchDatastore_Task, 'duration_secs': 0.009261} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.636697] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.636996] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 3401be45-c3ab-4780-81a7-a93b05742414/3401be45-c3ab-4780-81a7-a93b05742414.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 723.646902] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 0df161e9-4a66-432e-9366-60f231c53e3d/0df161e9-4a66-432e-9366-60f231c53e3d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 723.647218] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b89f49e-a708-4ec6-9555-7d6a1677082c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.651140] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6c7b951-eb16-4a18-bdfd-ee2aff1b3319 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.670680] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 723.670680] env[61768]: value = "task-1228639" [ 723.670680] env[61768]: _type = "Task" [ 723.670680] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.671941] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 723.671941] env[61768]: value = "task-1228640" [ 723.671941] env[61768]: _type = "Task" [ 723.671941] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.675536] env[61768]: DEBUG oslo_concurrency.lockutils [req-220fa071-884b-4175-825b-7057be6df3bd req-20d3af3f-8b32-4f05-a49c-aa9b349b2374 service nova] Releasing lock "refresh_cache-3401be45-c3ab-4780-81a7-a93b05742414" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.683039] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228639, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.685978] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228640, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.723460] env[61768]: DEBUG nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.805341] env[61768]: DEBUG oslo_concurrency.lockutils [req-1598e4c3-46e7-440d-985f-dc885392a6d9 req-52b55b50-712f-4d91-8154-f6f06142a9ce service nova] Releasing lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.805341] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquired lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.805341] env[61768]: DEBUG nova.network.neutron [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 724.113806] env[61768]: DEBUG nova.scheduler.client.report [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 724.185970] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228639, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.190855] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228640, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.555564] env[61768]: DEBUG nova.network.neutron [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updating instance_info_cache with network_info: [{"id": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "address": "fa:16:3e:c2:c0:a4", "network": {"id": "638c557a-bdca-4fa0-a3bb-ee6a1a5c4298", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1345341842-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76967cef64d8478ab599e0f336d67535", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "095fbf26-7367-4f9e-87c5-2965b64b0b0f", "external-id": "nsx-vlan-transportzone-777", "segmentation_id": 777, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42b4fa2d-89", "ovs_interfaceid": "42b4fa2d-897f-4316-85f8-b2168ccfabdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.620467] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.914s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.623397] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.740s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.646941] env[61768]: INFO nova.scheduler.client.report [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Deleted allocations for instance 63704075-5eaf-4f84-a90b-3a0a3e904a9d [ 724.683233] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228639, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.610369} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.683711] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 3401be45-c3ab-4780-81a7-a93b05742414/3401be45-c3ab-4780-81a7-a93b05742414.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 724.683984] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 724.684209] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a651bf2-7534-41cd-b034-587a7ca1a7c9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.689171] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228640, 'name': ReconfigVM_Task, 'duration_secs': 0.607018} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.690173] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 0df161e9-4a66-432e-9366-60f231c53e3d/0df161e9-4a66-432e-9366-60f231c53e3d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 724.691176] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-741a53bc-c361-44bd-8fb9-605d07b1e8c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.694645] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 724.694645] env[61768]: value = "task-1228641" [ 724.694645] env[61768]: _type = "Task" [ 724.694645] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.699220] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 724.699220] env[61768]: value = "task-1228642" [ 724.699220] env[61768]: _type = "Task" [ 724.699220] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.704868] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228641, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.709817] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228642, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.733840] env[61768]: DEBUG nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 724.760818] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 724.761328] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 724.761328] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 724.761610] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 724.761776] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 724.761943] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 724.762162] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 724.762362] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 724.762486] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 724.762674] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 724.762859] env[61768]: DEBUG nova.virt.hardware [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 724.764032] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3523b3a-19d1-4885-a0e5-59205f23687d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.771805] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef41210-30fd-401d-b6b1-0e2000f9512d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.059131] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Releasing lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.063117] env[61768]: DEBUG nova.compute.manager [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.064093] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c08ffb-c84c-44e0-83f7-e95678356bd4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.161095] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94ae9eea-ab1f-4e21-b982-3f0246e8a607 tempest-ServerShowV254Test-304766939 tempest-ServerShowV254Test-304766939-project-member] Lock "63704075-5eaf-4f84-a90b-3a0a3e904a9d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.999s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.207528] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228641, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064483} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.210669] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 725.211670] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023f6463-4694-4897-840d-0012a4fe9f18 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.218419] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228642, 'name': Rename_Task, 'duration_secs': 0.395655} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.218419] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 725.218563] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5e8d8e1-a1df-48e0-9f21-1968a52d8c07 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.237945] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 3401be45-c3ab-4780-81a7-a93b05742414/3401be45-c3ab-4780-81a7-a93b05742414.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 725.241624] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-790a607b-1841-414b-ab7d-fd2ca6093745 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.259551] env[61768]: DEBUG nova.compute.manager [req-1c61c8f8-bcf0-473e-8266-c6f3e4c7a969 req-8e219be0-a4e8-411c-a3b5-06ee27b89f7d service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Received event network-vif-plugged-932b7671-16df-463c-b31c-e99ede6a0287 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.259773] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c61c8f8-bcf0-473e-8266-c6f3e4c7a969 req-8e219be0-a4e8-411c-a3b5-06ee27b89f7d service nova] Acquiring lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.259981] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c61c8f8-bcf0-473e-8266-c6f3e4c7a969 req-8e219be0-a4e8-411c-a3b5-06ee27b89f7d service nova] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.260170] env[61768]: DEBUG oslo_concurrency.lockutils [req-1c61c8f8-bcf0-473e-8266-c6f3e4c7a969 req-8e219be0-a4e8-411c-a3b5-06ee27b89f7d service nova] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.260335] env[61768]: DEBUG nova.compute.manager [req-1c61c8f8-bcf0-473e-8266-c6f3e4c7a969 req-8e219be0-a4e8-411c-a3b5-06ee27b89f7d service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] No waiting events found dispatching network-vif-plugged-932b7671-16df-463c-b31c-e99ede6a0287 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 725.260508] env[61768]: WARNING nova.compute.manager [req-1c61c8f8-bcf0-473e-8266-c6f3e4c7a969 req-8e219be0-a4e8-411c-a3b5-06ee27b89f7d service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Received unexpected event network-vif-plugged-932b7671-16df-463c-b31c-e99ede6a0287 for instance with vm_state building and task_state spawning. [ 725.261126] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 725.261126] env[61768]: value = "task-1228643" [ 725.261126] env[61768]: _type = "Task" [ 725.261126] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.267486] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 725.267486] env[61768]: value = "task-1228644" [ 725.267486] env[61768]: _type = "Task" [ 725.267486] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.274270] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228643, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.281912] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.313563] env[61768]: DEBUG nova.network.neutron [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Successfully updated port: 932b7671-16df-463c-b31c-e99ede6a0287 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 725.516663] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-798f1eb0-b20b-41a2-9cdb-f71e1abb2cb6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.526195] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c5f835-0fa7-4394-8cc2-0f436918fa2b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.560522] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6da50cf-2b8b-4778-a2d5-60e66e8d5196 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.568779] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a013657e-0523-431f-acbe-f333a589f25d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.585167] env[61768]: DEBUG nova.compute.provider_tree [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.771758] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228643, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.779462] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228644, 'name': ReconfigVM_Task, 'duration_secs': 0.288549} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.779682] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 3401be45-c3ab-4780-81a7-a93b05742414/3401be45-c3ab-4780-81a7-a93b05742414.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 725.780279] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ddea7f8-c65b-430c-85e8-109188f8c58c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.786392] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 725.786392] env[61768]: value = "task-1228645" [ 725.786392] env[61768]: _type = "Task" [ 725.786392] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.794173] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228645, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.816056] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.816201] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquired lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.816357] env[61768]: DEBUG nova.network.neutron [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.833162] env[61768]: DEBUG nova.compute.manager [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.834340] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea79ab23-f10d-471d-ac2b-ccc3713f1314 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.089782] env[61768]: DEBUG nova.scheduler.client.report [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 726.093855] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d6feb3-5000-4c51-9641-88945037ee30 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.101434] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Doing hard reboot of VM {{(pid=61768) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 726.101683] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-89dc37e9-cb62-4af1-b531-675d97549e7e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.107529] env[61768]: DEBUG oslo_vmware.api [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 726.107529] env[61768]: value = "task-1228646" [ 726.107529] env[61768]: _type = "Task" [ 726.107529] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.117172] env[61768]: DEBUG oslo_vmware.api [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228646, 'name': ResetVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.273120] env[61768]: DEBUG oslo_vmware.api [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228643, 'name': PowerOnVM_Task, 'duration_secs': 0.829666} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.275009] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 726.275280] env[61768]: INFO nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Took 9.52 seconds to spawn the instance on the hypervisor. [ 726.275523] env[61768]: DEBUG nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 726.276346] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eae5bee-f575-4b2a-be42-d51c1443e3e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.294726] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228645, 'name': Rename_Task, 'duration_secs': 0.133473} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.295016] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 726.295356] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cdcab82b-4ece-427b-b44d-128791a229fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.304111] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 726.304111] env[61768]: value = "task-1228647" [ 726.304111] env[61768]: _type = "Task" [ 726.304111] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.317234] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228647, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.347139] env[61768]: INFO nova.compute.manager [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] instance snapshotting [ 726.347425] env[61768]: WARNING nova.compute.manager [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 726.349990] env[61768]: DEBUG nova.network.neutron [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.352929] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09fbb18-739b-4d83-846d-45003f877c41 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.375029] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b3318b-7e22-42db-8fe7-59bbf94f78f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.550268] env[61768]: DEBUG nova.network.neutron [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Updating instance_info_cache with network_info: [{"id": "932b7671-16df-463c-b31c-e99ede6a0287", "address": "fa:16:3e:fc:9e:48", "network": {"id": "385145e1-8386-4ab6-b9ae-bd5329b9bee6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-97019125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55c519433a0e4f2b85fec7f8ea820496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bc3fa06-9d5b-4ab1-8113-6ed8942d23b6", "external-id": "nsx-vlan-transportzone-72", "segmentation_id": 72, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932b7671-16", "ovs_interfaceid": "932b7671-16df-463c-b31c-e99ede6a0287", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.598384] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.976s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.599054] env[61768]: ERROR nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Traceback (most recent call last): [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self.driver.spawn(context, instance, image_meta, [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] vm_ref = self.build_virtual_machine(instance, [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.599054] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] for vif in network_info: [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return self._sync_wrapper(fn, *args, **kwargs) [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self.wait() [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self[:] = self._gt.wait() [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return self._exit_event.wait() [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] result = hub.switch() [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.599470] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return self.greenlet.switch() [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] result = function(*args, **kwargs) [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] return func(*args, **kwargs) [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] raise e [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] nwinfo = self.network_api.allocate_for_instance( [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] created_port_ids = self._update_ports_for_instance( [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] with excutils.save_and_reraise_exception(): [ 726.599902] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] self.force_reraise() [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] raise self.value [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] updated_port = self._update_port( [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] _ensure_no_port_binding_failure(port) [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] raise exception.PortBindingFailed(port_id=port['id']) [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] nova.exception.PortBindingFailed: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. [ 726.600464] env[61768]: ERROR nova.compute.manager [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] [ 726.600830] env[61768]: DEBUG nova.compute.utils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 726.601598] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.061s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.604557] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Build of instance a83f5b0f-508c-4829-a4a6-641baa7b4c95 was re-scheduled: Binding failed for port 729d32df-bf2d-4fc2-ae66-57fca1610887, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 726.605060] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 726.605299] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Acquiring lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.605462] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Acquired lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.605626] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.618536] env[61768]: DEBUG oslo_vmware.api [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228646, 'name': ResetVM_Task, 'duration_secs': 0.115993} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.618536] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Did hard reboot of VM {{(pid=61768) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 726.618536] env[61768]: DEBUG nova.compute.manager [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 726.619293] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c782311d-b8f7-4d48-8814-19d13dad896f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.794142] env[61768]: INFO nova.compute.manager [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Took 31.56 seconds to build instance. [ 726.814255] env[61768]: DEBUG oslo_vmware.api [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228647, 'name': PowerOnVM_Task, 'duration_secs': 0.501846} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.814502] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 726.814698] env[61768]: INFO nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Took 7.54 seconds to spawn the instance on the hypervisor. [ 726.814893] env[61768]: DEBUG nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 726.815650] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f8dff74-3b28-42fa-b44c-ed106224f2f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.885164] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 726.885463] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0b0d7219-0ba3-431d-b2bb-4ece7d330056 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.894882] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 726.894882] env[61768]: value = "task-1228648" [ 726.894882] env[61768]: _type = "Task" [ 726.894882] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.902487] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228648, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.053170] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Releasing lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.053478] env[61768]: DEBUG nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Instance network_info: |[{"id": "932b7671-16df-463c-b31c-e99ede6a0287", "address": "fa:16:3e:fc:9e:48", "network": {"id": "385145e1-8386-4ab6-b9ae-bd5329b9bee6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-97019125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55c519433a0e4f2b85fec7f8ea820496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bc3fa06-9d5b-4ab1-8113-6ed8942d23b6", "external-id": "nsx-vlan-transportzone-72", "segmentation_id": 72, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932b7671-16", "ovs_interfaceid": "932b7671-16df-463c-b31c-e99ede6a0287", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 727.053911] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:9e:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8bc3fa06-9d5b-4ab1-8113-6ed8942d23b6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '932b7671-16df-463c-b31c-e99ede6a0287', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 727.061781] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Creating folder: Project (55c519433a0e4f2b85fec7f8ea820496). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 727.062096] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2eef9dd7-ccd1-4dce-8f2b-a619a2952b87 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.073166] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Created folder: Project (55c519433a0e4f2b85fec7f8ea820496) in parent group-v265360. [ 727.073368] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Creating folder: Instances. Parent ref: group-v265393. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 727.073587] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5726460b-8363-4d7f-b88c-a983fa222632 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.081690] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Created folder: Instances in parent group-v265393. [ 727.081936] env[61768]: DEBUG oslo.service.loopingcall [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.082143] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 727.082340] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8c625f1-d3a4-4beb-8425-91a6b9148caa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.100899] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 727.100899] env[61768]: value = "task-1228651" [ 727.100899] env[61768]: _type = "Task" [ 727.100899] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.113439] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228651, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.127171] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.131716] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f600fec6-cedb-45de-9205-3ddb6da4b219 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.106s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.185823] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.298998] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1f90f738-9cce-49b5-bce8-9b3d54119779 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "0df161e9-4a66-432e-9366-60f231c53e3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.895s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.335860] env[61768]: INFO nova.compute.manager [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Took 30.23 seconds to build instance. [ 727.405614] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228648, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.421675] env[61768]: DEBUG nova.compute.manager [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Received event network-changed-932b7671-16df-463c-b31c-e99ede6a0287 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.421950] env[61768]: DEBUG nova.compute.manager [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Refreshing instance network info cache due to event network-changed-932b7671-16df-463c-b31c-e99ede6a0287. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 727.422150] env[61768]: DEBUG oslo_concurrency.lockutils [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] Acquiring lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.422292] env[61768]: DEBUG oslo_concurrency.lockutils [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] Acquired lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.422455] env[61768]: DEBUG nova.network.neutron [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Refreshing network info cache for port 932b7671-16df-463c-b31c-e99ede6a0287 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 727.441443] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdcda117-699b-425a-80b0-d1570be11a66 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.449290] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3f0d43c-0f11-44fb-87d2-3213eb9756e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.480901] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98719666-12a5-43f8-85f6-27fd3d1fca65 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.488332] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9930a549-5e71-4240-b81d-0513a66cd518 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.501555] env[61768]: DEBUG nova.compute.provider_tree [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.611091] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228651, 'name': CreateVM_Task, 'duration_secs': 0.312158} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.611312] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 727.612090] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.612284] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.612646] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 727.612956] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9bbefa32-68fd-4b3c-b885-f3e5b16c8417 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.617756] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 727.617756] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a0ebc6-a790-91e5-86a4-821b02ea80ce" [ 727.617756] env[61768]: _type = "Task" [ 727.617756] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.625579] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a0ebc6-a790-91e5-86a4-821b02ea80ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.689222] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Releasing lock "refresh_cache-a83f5b0f-508c-4829-a4a6-641baa7b4c95" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.689503] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 727.689681] env[61768]: DEBUG nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.689857] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 727.715059] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.802225] env[61768]: DEBUG nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.838156] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80f0d2bf-ba35-4b1f-bc33-9418a1e5e4f2 tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "3401be45-c3ab-4780-81a7-a93b05742414" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.911s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.907963] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228648, 'name': CreateSnapshot_Task, 'duration_secs': 0.591753} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.908294] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 727.909019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bd416ed-210f-4f47-9fe0-76b393d60d62 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.008455] env[61768]: DEBUG nova.scheduler.client.report [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.128772] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a0ebc6-a790-91e5-86a4-821b02ea80ce, 'name': SearchDatastore_Task, 'duration_secs': 0.009196} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.129518] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.129942] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 728.130321] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.130636] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.131012] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 728.133457] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d1c9987-ec74-48fa-98f5-05c17e9dea23 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.143019] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 728.143019] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 728.143019] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-46248825-fef0-4d1e-9ee3-67cac74edc80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.146955] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 728.146955] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b6c7dd-89fa-94bb-049b-9b5acc19a81b" [ 728.146955] env[61768]: _type = "Task" [ 728.146955] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.155571] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b6c7dd-89fa-94bb-049b-9b5acc19a81b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.211346] env[61768]: DEBUG nova.network.neutron [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Updated VIF entry in instance network info cache for port 932b7671-16df-463c-b31c-e99ede6a0287. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 728.211700] env[61768]: DEBUG nova.network.neutron [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Updating instance_info_cache with network_info: [{"id": "932b7671-16df-463c-b31c-e99ede6a0287", "address": "fa:16:3e:fc:9e:48", "network": {"id": "385145e1-8386-4ab6-b9ae-bd5329b9bee6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-97019125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55c519433a0e4f2b85fec7f8ea820496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bc3fa06-9d5b-4ab1-8113-6ed8942d23b6", "external-id": "nsx-vlan-transportzone-72", "segmentation_id": 72, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932b7671-16", "ovs_interfaceid": "932b7671-16df-463c-b31c-e99ede6a0287", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.218411] env[61768]: DEBUG nova.network.neutron [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.327960] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.341140] env[61768]: DEBUG nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 728.427405] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 728.427405] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0c0f1419-9b66-4c42-bbd0-f76b2ae84459 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.435874] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 728.435874] env[61768]: value = "task-1228652" [ 728.435874] env[61768]: _type = "Task" [ 728.435874] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.444277] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228652, 'name': CloneVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.513599] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.912s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.514376] env[61768]: ERROR nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Traceback (most recent call last): [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self.driver.spawn(context, instance, image_meta, [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] vm_ref = self.build_virtual_machine(instance, [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.514376] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] for vif in network_info: [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return self._sync_wrapper(fn, *args, **kwargs) [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self.wait() [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self[:] = self._gt.wait() [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return self._exit_event.wait() [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] result = hub.switch() [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.514853] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return self.greenlet.switch() [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] result = function(*args, **kwargs) [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] return func(*args, **kwargs) [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] raise e [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] nwinfo = self.network_api.allocate_for_instance( [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] created_port_ids = self._update_ports_for_instance( [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] with excutils.save_and_reraise_exception(): [ 728.515318] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] self.force_reraise() [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] raise self.value [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] updated_port = self._update_port( [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] _ensure_no_port_binding_failure(port) [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] raise exception.PortBindingFailed(port_id=port['id']) [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] nova.exception.PortBindingFailed: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. [ 728.515716] env[61768]: ERROR nova.compute.manager [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] [ 728.516151] env[61768]: DEBUG nova.compute.utils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.516581] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.672s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.518303] env[61768]: INFO nova.compute.claims [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.522339] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Build of instance f6b07607-632d-46d9-a72f-5a524fd20a3b was re-scheduled: Binding failed for port c4c98ad0-6b0f-4a0e-af7f-581439418979, please check neutron logs for more information. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.522826] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Unplugging VIFs for instance {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.523096] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Acquiring lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.523224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Acquired lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.523396] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 728.661773] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b6c7dd-89fa-94bb-049b-9b5acc19a81b, 'name': SearchDatastore_Task, 'duration_secs': 0.021328} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.662946] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6990719f-e896-4c5c-acdf-47ab30e1b1c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.669632] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 728.669632] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a9b13-89d4-4746-08c3-ab7e948e27bd" [ 728.669632] env[61768]: _type = "Task" [ 728.669632] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.680167] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a9b13-89d4-4746-08c3-ab7e948e27bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.698464] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.698816] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.699139] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.699424] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.699685] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.702586] env[61768]: INFO nova.compute.manager [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Terminating instance [ 728.707934] env[61768]: DEBUG nova.compute.manager [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.708179] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 728.709016] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4940e078-df73-4690-bcbb-503d632655c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.717223] env[61768]: DEBUG oslo_concurrency.lockutils [req-4722d150-a8bf-451f-b888-3acede00f6b4 req-2fa1a1a5-ba02-48de-ace3-853ae2857d7e service nova] Releasing lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.717603] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 728.717861] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-176e7d2f-f2f3-4c9c-bd29-a3e44eecb7ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.723038] env[61768]: INFO nova.compute.manager [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] [instance: a83f5b0f-508c-4829-a4a6-641baa7b4c95] Took 1.03 seconds to deallocate network for instance. [ 728.726652] env[61768]: DEBUG oslo_vmware.api [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 728.726652] env[61768]: value = "task-1228653" [ 728.726652] env[61768]: _type = "Task" [ 728.726652] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.735127] env[61768]: DEBUG oslo_vmware.api [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228653, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.867156] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.949529] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228652, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.049900] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.166361] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.180899] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a9b13-89d4-4746-08c3-ab7e948e27bd, 'name': SearchDatastore_Task, 'duration_secs': 0.024937} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.181888] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.182289] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 91d56a8c-0664-4b3c-bf28-e2c668ac65fa/91d56a8c-0664-4b3c-bf28-e2c668ac65fa.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 729.182620] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f586afd0-a929-48cb-9646-70ac4d6bdba4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.190084] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 729.190084] env[61768]: value = "task-1228654" [ 729.190084] env[61768]: _type = "Task" [ 729.190084] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.200195] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228654, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.240942] env[61768]: DEBUG oslo_vmware.api [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228653, 'name': PowerOffVM_Task, 'duration_secs': 0.248914} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.241284] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 729.241462] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 729.241879] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-586e8d4e-bfb4-4f98-8e2d-aacdf4f73148 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.308699] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 729.308699] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 729.308699] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Deleting the datastore file [datastore2] 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 729.308699] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f372a3ce-d53b-4876-80ca-495088ef0347 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.315714] env[61768]: DEBUG oslo_vmware.api [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for the task: (returnval){ [ 729.315714] env[61768]: value = "task-1228656" [ 729.315714] env[61768]: _type = "Task" [ 729.315714] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.323573] env[61768]: DEBUG oslo_vmware.api [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228656, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.449093] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228652, 'name': CloneVM_Task} progress is 95%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.673581] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Releasing lock "refresh_cache-f6b07607-632d-46d9-a72f-5a524fd20a3b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.673752] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61768) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.673929] env[61768]: DEBUG nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.674110] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 729.691186] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.703784] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228654, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500884} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.704900] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 91d56a8c-0664-4b3c-bf28-e2c668ac65fa/91d56a8c-0664-4b3c-bf28-e2c668ac65fa.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 729.705138] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 729.708120] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e4b20f5c-40a7-49f5-8854-b9cd0ecee13b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.716025] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 729.716025] env[61768]: value = "task-1228657" [ 729.716025] env[61768]: _type = "Task" [ 729.716025] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.723887] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228657, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.733188] env[61768]: DEBUG nova.compute.manager [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Received event network-changed-42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.733188] env[61768]: DEBUG nova.compute.manager [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Refreshing instance network info cache due to event network-changed-42b4fa2d-897f-4316-85f8-b2168ccfabdb. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.733188] env[61768]: DEBUG oslo_concurrency.lockutils [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] Acquiring lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.733188] env[61768]: DEBUG oslo_concurrency.lockutils [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] Acquired lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.733188] env[61768]: DEBUG nova.network.neutron [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Refreshing network info cache for port 42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 729.763749] env[61768]: INFO nova.scheduler.client.report [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Deleted allocations for instance a83f5b0f-508c-4829-a4a6-641baa7b4c95 [ 729.831368] env[61768]: DEBUG oslo_vmware.api [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Task: {'id': task-1228656, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.329725} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.831990] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 729.831990] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 729.832110] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 729.832248] env[61768]: INFO nova.compute.manager [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Took 1.12 seconds to destroy the instance on the hypervisor. [ 729.832633] env[61768]: DEBUG oslo.service.loopingcall [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.832706] env[61768]: DEBUG nova.compute.manager [-] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.832784] env[61768]: DEBUG nova.network.neutron [-] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 729.862681] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5063dfb8-efbe-4766-9c5f-dfd0d7ba1cb6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.876542] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86581abe-3803-484a-ae63-b2c97d486c0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.909630] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "3401be45-c3ab-4780-81a7-a93b05742414" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.909920] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "3401be45-c3ab-4780-81a7-a93b05742414" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.910156] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "3401be45-c3ab-4780-81a7-a93b05742414-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.910346] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "3401be45-c3ab-4780-81a7-a93b05742414-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.910738] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "3401be45-c3ab-4780-81a7-a93b05742414-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.914218] env[61768]: INFO nova.compute.manager [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Terminating instance [ 729.916761] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e72e4ca-3602-4a11-96af-c08872bfdbfd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.920888] env[61768]: DEBUG nova.compute.manager [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.921140] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 729.922267] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebdb3f7-77cc-4d40-9b09-d325bed5abcc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.932797] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab58abdb-450b-4b21-a5f2-12681af15e2f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.936558] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 729.936803] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0d11ea1a-da05-4a37-a595-e08b60e8a901 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.950304] env[61768]: DEBUG nova.compute.provider_tree [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.953724] env[61768]: DEBUG oslo_vmware.api [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 729.953724] env[61768]: value = "task-1228658" [ 729.953724] env[61768]: _type = "Task" [ 729.953724] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.960586] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228652, 'name': CloneVM_Task, 'duration_secs': 1.241333} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.961293] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Created linked-clone VM from snapshot [ 729.962068] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fbbd83-2d70-4c97-ab2e-5e8565e4f98c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.968978] env[61768]: DEBUG oslo_vmware.api [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228658, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.974238] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Uploading image c5b4c304-600a-43d8-802c-259ae58cbf74 {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 730.005127] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 730.005127] env[61768]: value = "vm-265397" [ 730.005127] env[61768]: _type = "VirtualMachine" [ 730.005127] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 730.007317] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3c2ca21b-7a3c-42fc-9289-08ea81e305c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.014364] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lease: (returnval){ [ 730.014364] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223c41b-0204-3a90-e08d-e55a7f9c8eef" [ 730.014364] env[61768]: _type = "HttpNfcLease" [ 730.014364] env[61768]: } obtained for exporting VM: (result){ [ 730.014364] env[61768]: value = "vm-265397" [ 730.014364] env[61768]: _type = "VirtualMachine" [ 730.014364] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 730.014587] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the lease: (returnval){ [ 730.014587] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223c41b-0204-3a90-e08d-e55a7f9c8eef" [ 730.014587] env[61768]: _type = "HttpNfcLease" [ 730.014587] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 730.022943] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 730.022943] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223c41b-0204-3a90-e08d-e55a7f9c8eef" [ 730.022943] env[61768]: _type = "HttpNfcLease" [ 730.022943] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 730.197636] env[61768]: DEBUG nova.network.neutron [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.225326] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228657, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.107471} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.225629] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 730.226542] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b53371c-1ec5-48a2-9978-68fbbca7b11f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.252488] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 91d56a8c-0664-4b3c-bf28-e2c668ac65fa/91d56a8c-0664-4b3c-bf28-e2c668ac65fa.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 730.253101] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39ccd111-0207-485c-82e4-192252d7107a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.273645] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a2887ccf-129e-4304-a538-bcd4d96f15cd tempest-ServerDiagnosticsNegativeTest-66140769 tempest-ServerDiagnosticsNegativeTest-66140769-project-member] Lock "a83f5b0f-508c-4829-a4a6-641baa7b4c95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 163.448s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.276254] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 730.276254] env[61768]: value = "task-1228660" [ 730.276254] env[61768]: _type = "Task" [ 730.276254] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.285875] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228660, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.290568] env[61768]: INFO nova.network.neutron [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Port 42b4fa2d-897f-4316-85f8-b2168ccfabdb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 730.290722] env[61768]: DEBUG nova.network.neutron [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.455219] env[61768]: DEBUG nova.scheduler.client.report [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.467476] env[61768]: DEBUG oslo_vmware.api [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228658, 'name': PowerOffVM_Task, 'duration_secs': 0.235689} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.467700] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 730.467869] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 730.468131] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f7bff5ea-145b-4f2b-9b01-6dc5cc2d3b4c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.523378] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 730.523378] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223c41b-0204-3a90-e08d-e55a7f9c8eef" [ 730.523378] env[61768]: _type = "HttpNfcLease" [ 730.523378] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 730.523378] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 730.523378] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223c41b-0204-3a90-e08d-e55a7f9c8eef" [ 730.523378] env[61768]: _type = "HttpNfcLease" [ 730.523378] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 730.523761] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40c1faf-d38d-4413-af1b-93d537d11aee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.535639] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5214d399-a27b-f27e-a51f-fbf3d963bfe4/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 730.535891] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5214d399-a27b-f27e-a51f-fbf3d963bfe4/disk-0.vmdk for reading. {{(pid=61768) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 730.539074] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 730.539074] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 730.539074] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Deleting the datastore file [datastore1] 3401be45-c3ab-4780-81a7-a93b05742414 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 730.540092] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bfec3567-ab1c-4308-a418-b238c0590d86 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.600383] env[61768]: DEBUG oslo_vmware.api [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for the task: (returnval){ [ 730.600383] env[61768]: value = "task-1228662" [ 730.600383] env[61768]: _type = "Task" [ 730.600383] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.609380] env[61768]: DEBUG oslo_vmware.api [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228662, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.620832] env[61768]: DEBUG nova.network.neutron [-] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.662830] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-417cc8bf-c705-45d7-b821-93a01d48f91b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.700194] env[61768]: INFO nova.compute.manager [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] [instance: f6b07607-632d-46d9-a72f-5a524fd20a3b] Took 1.03 seconds to deallocate network for instance. [ 730.780965] env[61768]: DEBUG nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.790926] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228660, 'name': ReconfigVM_Task, 'duration_secs': 0.256409} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.791203] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 91d56a8c-0664-4b3c-bf28-e2c668ac65fa/91d56a8c-0664-4b3c-bf28-e2c668ac65fa.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.791782] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d306dae9-f766-478e-b161-257239a57f02 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.793573] env[61768]: DEBUG oslo_concurrency.lockutils [req-e376f3a8-ef6f-439a-bdd2-d0e62433d1c1 req-2a9579e7-6b99-42d0-992b-55078e4ee86e service nova] Releasing lock "refresh_cache-7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.798922] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 730.798922] env[61768]: value = "task-1228663" [ 730.798922] env[61768]: _type = "Task" [ 730.798922] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.806396] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228663, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.837431] env[61768]: DEBUG nova.compute.manager [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 730.838663] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43680cae-e271-4ee4-b531-493523d7a6c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.966898] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.971194] env[61768]: DEBUG nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.974767] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.206s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.977618] env[61768]: INFO nova.compute.claims [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.114036] env[61768]: DEBUG oslo_vmware.api [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Task: {'id': task-1228662, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.2008} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.114673] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 731.115990] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 731.115990] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 731.115990] env[61768]: INFO nova.compute.manager [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Took 1.19 seconds to destroy the instance on the hypervisor. [ 731.115990] env[61768]: DEBUG oslo.service.loopingcall [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.115990] env[61768]: DEBUG nova.compute.manager [-] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.116442] env[61768]: DEBUG nova.network.neutron [-] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 731.122801] env[61768]: INFO nova.compute.manager [-] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Took 1.29 seconds to deallocate network for instance. [ 731.310461] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228663, 'name': Rename_Task, 'duration_secs': 0.151569} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.310781] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 731.311076] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e25a83f-2e43-4b92-ae10-89b95c54e6c9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.313724] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.319234] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 731.319234] env[61768]: value = "task-1228664" [ 731.319234] env[61768]: _type = "Task" [ 731.319234] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.328433] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228664, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.352235] env[61768]: INFO nova.compute.manager [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] instance snapshotting [ 731.356684] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b08460e-5123-49a5-9a13-9f8e07c6efdc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.380672] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a075fa-4f7f-4c8a-bd2b-ed65eac6bbee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.483451] env[61768]: DEBUG nova.compute.utils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.486938] env[61768]: DEBUG nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.487183] env[61768]: DEBUG nova.network.neutron [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 731.533121] env[61768]: DEBUG nova.policy [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9d7f72b4c22452d89004ca70d165e67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f4804c086b449608ef34ea22c018907', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.633043] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.745323] env[61768]: INFO nova.scheduler.client.report [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Deleted allocations for instance f6b07607-632d-46d9-a72f-5a524fd20a3b [ 731.777299] env[61768]: DEBUG nova.compute.manager [req-145673c6-8edb-4d1d-9dc7-4d5440d48db0 req-2d52b980-ca9b-44ee-a078-93b2b71de644 service nova] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Received event network-vif-deleted-42b4fa2d-897f-4316-85f8-b2168ccfabdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.778619] env[61768]: DEBUG nova.compute.manager [req-145673c6-8edb-4d1d-9dc7-4d5440d48db0 req-2d52b980-ca9b-44ee-a078-93b2b71de644 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Received event network-vif-deleted-90ce2022-9af9-4df8-b92c-140069264092 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.779563] env[61768]: INFO nova.compute.manager [req-145673c6-8edb-4d1d-9dc7-4d5440d48db0 req-2d52b980-ca9b-44ee-a078-93b2b71de644 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Neutron deleted interface 90ce2022-9af9-4df8-b92c-140069264092; detaching it from the instance and deleting it from the info cache [ 731.779955] env[61768]: DEBUG nova.network.neutron [req-145673c6-8edb-4d1d-9dc7-4d5440d48db0 req-2d52b980-ca9b-44ee-a078-93b2b71de644 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.830403] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228664, 'name': PowerOnVM_Task} progress is 83%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.843607] env[61768]: DEBUG nova.network.neutron [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Successfully created port: 5379c85e-3ce3-433a-ad59-3b790a40b5fb {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 731.895679] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 731.895960] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-362ddf05-ec24-45bb-8321-f6897e7aeeb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.905454] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 731.905454] env[61768]: value = "task-1228665" [ 731.905454] env[61768]: _type = "Task" [ 731.905454] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.916920] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228665, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.977940] env[61768]: DEBUG nova.network.neutron [-] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.987858] env[61768]: DEBUG nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.259066] env[61768]: DEBUG oslo_concurrency.lockutils [None req-63f2c37b-f626-40b8-bb7e-b3dae4c61bfd tempest-ServerActionsTestOtherB-78416712 tempest-ServerActionsTestOtherB-78416712-project-member] Lock "f6b07607-632d-46d9-a72f-5a524fd20a3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 164.454s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.283811] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0f4e2d9e-05ba-431b-a1e0-453c804f6466 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.305401] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401faa0f-1201-4674-9b30-47cd871b884c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.336394] env[61768]: DEBUG oslo_vmware.api [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228664, 'name': PowerOnVM_Task, 'duration_secs': 0.725353} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.350631] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 732.350974] env[61768]: INFO nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Took 7.62 seconds to spawn the instance on the hypervisor. [ 732.351242] env[61768]: DEBUG nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 732.352354] env[61768]: DEBUG nova.compute.manager [req-145673c6-8edb-4d1d-9dc7-4d5440d48db0 req-2d52b980-ca9b-44ee-a078-93b2b71de644 service nova] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Detach interface failed, port_id=90ce2022-9af9-4df8-b92c-140069264092, reason: Instance 3401be45-c3ab-4780-81a7-a93b05742414 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 732.353625] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3678342d-af3a-43bf-9ba7-79455c5a15fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.375661] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4adc62f7-f524-44e0-9345-eb551b612d77 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.384025] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-922d98ff-f2eb-45a9-bdad-79e536dea1ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.425831] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e43841-ca45-4b61-89c9-b849b8e2920b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.435363] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228665, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.438781] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be65721-3462-4f0b-8058-d8e32ef4e275 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.454547] env[61768]: DEBUG nova.compute.provider_tree [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.482373] env[61768]: INFO nova.compute.manager [-] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Took 1.37 seconds to deallocate network for instance. [ 732.762928] env[61768]: DEBUG nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.874834] env[61768]: INFO nova.compute.manager [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Took 31.74 seconds to build instance. [ 732.934221] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228665, 'name': CreateSnapshot_Task, 'duration_secs': 0.658718} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.934541] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 732.936816] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f03146-5d05-4c50-a4e0-a672fdda3bc9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.957449] env[61768]: DEBUG nova.scheduler.client.report [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.990051] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.004544] env[61768]: DEBUG nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 733.049991] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.049991] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.053028] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.053028] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.053028] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.053028] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.053028] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.053313] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.053313] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.053313] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.053313] env[61768]: DEBUG nova.virt.hardware [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.053618] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8029f9-9b01-4804-8af8-47e644859cc5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.062620] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a332adee-376a-4cde-bddd-eeb76224d2f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.289822] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.332033] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.381107] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fa92855c-e392-41fc-844c-7e1a491cfe33 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.465s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.381107] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.048s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.381107] env[61768]: INFO nova.compute.manager [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Rebooting instance [ 733.458488] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 733.461431] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a5ac3d09-6912-4a30-b4cc-c141972d17e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.464212] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.466972] env[61768]: DEBUG nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.468440] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 23.274s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.468985] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.469362] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 733.473187] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.971s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.473187] env[61768]: INFO nova.compute.claims [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.475851] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1cdd2d-1956-4db2-bc80-53734b1db163 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.482459] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 733.482459] env[61768]: value = "task-1228666" [ 733.482459] env[61768]: _type = "Task" [ 733.482459] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.490315] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae0f2c5e-3742-4a0d-ba0c-428f641f0294 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.500338] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228666, 'name': CloneVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.509488] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-008d11ac-7a3b-48d1-8521-f155593f492d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.517344] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86cb5e68-6aad-4894-9809-456a2345d080 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.558917] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181251MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 733.559590] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.573701] env[61768]: DEBUG nova.network.neutron [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Successfully updated port: 5379c85e-3ce3-433a-ad59-3b790a40b5fb {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 733.809353] env[61768]: DEBUG nova.compute.manager [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Received event network-vif-plugged-5379c85e-3ce3-433a-ad59-3b790a40b5fb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.809601] env[61768]: DEBUG oslo_concurrency.lockutils [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] Acquiring lock "efbde802-8cb0-4563-a776-3722a3889afe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.809825] env[61768]: DEBUG oslo_concurrency.lockutils [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] Lock "efbde802-8cb0-4563-a776-3722a3889afe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.810217] env[61768]: DEBUG oslo_concurrency.lockutils [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] Lock "efbde802-8cb0-4563-a776-3722a3889afe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.810646] env[61768]: DEBUG nova.compute.manager [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] No waiting events found dispatching network-vif-plugged-5379c85e-3ce3-433a-ad59-3b790a40b5fb {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 733.810646] env[61768]: WARNING nova.compute.manager [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Received unexpected event network-vif-plugged-5379c85e-3ce3-433a-ad59-3b790a40b5fb for instance with vm_state building and task_state spawning. [ 733.810754] env[61768]: DEBUG nova.compute.manager [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Received event network-changed-5379c85e-3ce3-433a-ad59-3b790a40b5fb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.811121] env[61768]: DEBUG nova.compute.manager [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Refreshing instance network info cache due to event network-changed-5379c85e-3ce3-433a-ad59-3b790a40b5fb. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 733.811351] env[61768]: DEBUG oslo_concurrency.lockutils [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] Acquiring lock "refresh_cache-efbde802-8cb0-4563-a776-3722a3889afe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.811502] env[61768]: DEBUG oslo_concurrency.lockutils [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] Acquired lock "refresh_cache-efbde802-8cb0-4563-a776-3722a3889afe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.811668] env[61768]: DEBUG nova.network.neutron [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Refreshing network info cache for port 5379c85e-3ce3-433a-ad59-3b790a40b5fb {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 733.881915] env[61768]: DEBUG nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 733.909373] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.909580] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquired lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.909762] env[61768]: DEBUG nova.network.neutron [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.978677] env[61768]: DEBUG nova.compute.utils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.981241] env[61768]: DEBUG nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.981438] env[61768]: DEBUG nova.network.neutron [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 733.997722] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228666, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.057559] env[61768]: DEBUG nova.policy [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d4d1da10adb473781e31865fa759631', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f446e9095b274a9abe29f764ff520164', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.077842] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-efbde802-8cb0-4563-a776-3722a3889afe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.377101] env[61768]: DEBUG nova.network.neutron [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.415720] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.482072] env[61768]: DEBUG nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.503605] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228666, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.528785] env[61768]: DEBUG nova.network.neutron [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.592786] env[61768]: DEBUG nova.network.neutron [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Successfully created port: 79d4c508-cec4-40c8-8974-b1916ac2a4d2 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 734.908503] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00db1904-b6a6-49a3-ae1a-c6c5f4163d8e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.918888] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cdcbce1-5e8f-4fa0-b7bc-037b6d9926b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.962404] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b2e3b95-44ca-4114-a630-86e90a47c623 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.972852] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367ad434-d97e-4284-bb0a-c699558820dc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.991739] env[61768]: DEBUG nova.compute.provider_tree [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.018772] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228666, 'name': CloneVM_Task} progress is 95%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.030139] env[61768]: DEBUG oslo_concurrency.lockutils [req-6489f318-c4cf-46a2-b08b-10990c644c5e req-8003972a-6292-43e7-96b5-36c26b838676 service nova] Releasing lock "refresh_cache-efbde802-8cb0-4563-a776-3722a3889afe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.030528] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-efbde802-8cb0-4563-a776-3722a3889afe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.030796] env[61768]: DEBUG nova.network.neutron [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 735.109311] env[61768]: DEBUG nova.network.neutron [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Updating instance_info_cache with network_info: [{"id": "932b7671-16df-463c-b31c-e99ede6a0287", "address": "fa:16:3e:fc:9e:48", "network": {"id": "385145e1-8386-4ab6-b9ae-bd5329b9bee6", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-97019125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55c519433a0e4f2b85fec7f8ea820496", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bc3fa06-9d5b-4ab1-8113-6ed8942d23b6", "external-id": "nsx-vlan-transportzone-72", "segmentation_id": 72, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932b7671-16", "ovs_interfaceid": "932b7671-16df-463c-b31c-e99ede6a0287", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.505499] env[61768]: DEBUG nova.scheduler.client.report [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.514039] env[61768]: DEBUG nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.520730] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228666, 'name': CloneVM_Task, 'duration_secs': 1.706565} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.520730] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Created linked-clone VM from snapshot [ 735.521606] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa415d4-e76e-4e01-b2fd-1514cb8be0a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.540390] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Uploading image 41d5ccba-d2fa-4a1f-8ebb-5cc3c3f1fb7e {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 735.552501] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.552818] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.553018] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.553854] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.553854] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.553854] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.554036] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.554381] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.554381] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.556870] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.556870] env[61768]: DEBUG nova.virt.hardware [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.556870] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1594890-08f1-4a1a-83e9-4dc2cb9f9474 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.560523] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 735.561398] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b6661477-a85b-44fb-b11a-0a020ce8a4ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.571183] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b03f4fc-964c-4b7a-8b6e-e05d223b6eef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.575963] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 735.575963] env[61768]: value = "task-1228667" [ 735.575963] env[61768]: _type = "Task" [ 735.575963] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.601742] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228667, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.605657] env[61768]: DEBUG nova.network.neutron [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.612019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Releasing lock "refresh_cache-91d56a8c-0664-4b3c-bf28-e2c668ac65fa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.614545] env[61768]: DEBUG nova.compute.manager [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 735.615493] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a632514e-8cf1-4219-91b0-51aaf2b005ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.848020] env[61768]: DEBUG nova.network.neutron [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Updating instance_info_cache with network_info: [{"id": "5379c85e-3ce3-433a-ad59-3b790a40b5fb", "address": "fa:16:3e:c2:02:09", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5379c85e-3c", "ovs_interfaceid": "5379c85e-3ce3-433a-ad59-3b790a40b5fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.016266] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.546s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.016848] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 736.025215] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.782s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.025666] env[61768]: INFO nova.compute.claims [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.092881] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228667, 'name': Destroy_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.350537] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-efbde802-8cb0-4563-a776-3722a3889afe" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.350537] env[61768]: DEBUG nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance network_info: |[{"id": "5379c85e-3ce3-433a-ad59-3b790a40b5fb", "address": "fa:16:3e:c2:02:09", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5379c85e-3c", "ovs_interfaceid": "5379c85e-3ce3-433a-ad59-3b790a40b5fb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 736.350955] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:02:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5379c85e-3ce3-433a-ad59-3b790a40b5fb', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.361814] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Creating folder: Project (9f4804c086b449608ef34ea22c018907). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 736.364230] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-947d5d54-c535-4281-96c4-f7a9abbff579 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.372921] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Created folder: Project (9f4804c086b449608ef34ea22c018907) in parent group-v265360. [ 736.373188] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Creating folder: Instances. Parent ref: group-v265400. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 736.373503] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-03fe14b5-af9e-4634-84b6-24de00773377 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.384417] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Created folder: Instances in parent group-v265400. [ 736.384729] env[61768]: DEBUG oslo.service.loopingcall [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.385682] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 736.385949] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2356a77-4258-4e90-bfc0-485351b8af6c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.411395] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 736.411395] env[61768]: value = "task-1228670" [ 736.411395] env[61768]: _type = "Task" [ 736.411395] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.421055] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228670, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.476467] env[61768]: DEBUG nova.network.neutron [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Successfully updated port: 79d4c508-cec4-40c8-8974-b1916ac2a4d2 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 736.529080] env[61768]: DEBUG nova.compute.utils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.529699] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.530072] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 736.557444] env[61768]: DEBUG nova.compute.manager [req-32208234-f1c5-44c7-96f7-6fb29412e41e req-50883769-32f9-4151-9240-2cfb85bc2b54 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Received event network-vif-plugged-79d4c508-cec4-40c8-8974-b1916ac2a4d2 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 736.557675] env[61768]: DEBUG oslo_concurrency.lockutils [req-32208234-f1c5-44c7-96f7-6fb29412e41e req-50883769-32f9-4151-9240-2cfb85bc2b54 service nova] Acquiring lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.557883] env[61768]: DEBUG oslo_concurrency.lockutils [req-32208234-f1c5-44c7-96f7-6fb29412e41e req-50883769-32f9-4151-9240-2cfb85bc2b54 service nova] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.558168] env[61768]: DEBUG oslo_concurrency.lockutils [req-32208234-f1c5-44c7-96f7-6fb29412e41e req-50883769-32f9-4151-9240-2cfb85bc2b54 service nova] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.558359] env[61768]: DEBUG nova.compute.manager [req-32208234-f1c5-44c7-96f7-6fb29412e41e req-50883769-32f9-4151-9240-2cfb85bc2b54 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] No waiting events found dispatching network-vif-plugged-79d4c508-cec4-40c8-8974-b1916ac2a4d2 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 736.558527] env[61768]: WARNING nova.compute.manager [req-32208234-f1c5-44c7-96f7-6fb29412e41e req-50883769-32f9-4151-9240-2cfb85bc2b54 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Received unexpected event network-vif-plugged-79d4c508-cec4-40c8-8974-b1916ac2a4d2 for instance with vm_state building and task_state spawning. [ 736.586912] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228667, 'name': Destroy_Task, 'duration_secs': 0.650099} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.589535] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Destroyed the VM [ 736.589535] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 736.589882] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2193a945-8267-45ea-a1c4-e55735348d2c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.600756] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 736.600756] env[61768]: value = "task-1228671" [ 736.600756] env[61768]: _type = "Task" [ 736.600756] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.607018] env[61768]: DEBUG nova.policy [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1600e5a4cc4e44cb8f6d384d6fa1028a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54b7f349107b4f56960cae84edbea9e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.615322] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228671, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.637010] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5624cf8e-efad-4f16-bb22-b23d6d119ed9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.647724] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Doing hard reboot of VM {{(pid=61768) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 736.648053] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-4254b6b1-244f-4f56-9258-48b1df13b6a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.656583] env[61768]: DEBUG oslo_vmware.api [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 736.656583] env[61768]: value = "task-1228672" [ 736.656583] env[61768]: _type = "Task" [ 736.656583] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.666762] env[61768]: DEBUG oslo_vmware.api [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228672, 'name': ResetVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.926354] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228670, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.981813] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "refresh_cache-7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.981813] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquired lock "refresh_cache-7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.981813] env[61768]: DEBUG nova.network.neutron [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 737.030714] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 737.112779] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228671, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.116664] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Successfully created port: 3c8dc29a-e757-4d7d-93a4-9274629ec62c {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 737.171575] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.171830] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.172060] env[61768]: DEBUG oslo_vmware.api [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228672, 'name': ResetVM_Task, 'duration_secs': 0.115768} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.172302] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Did hard reboot of VM {{(pid=61768) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 737.172542] env[61768]: DEBUG nova.compute.manager [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 737.173502] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856b066a-32bd-406c-8b79-2a013a9ff197 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.360931] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3acd168e-575b-4f18-aec6-65f4a9ee4e8d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.369171] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327e9384-c064-4521-b372-3d530e7029f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.399328] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744587b0-251e-4284-a063-bc4ba85f3b9e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.406631] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd90d34b-fe09-41d5-be30-1e37ce866c95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.420144] env[61768]: DEBUG nova.compute.provider_tree [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.428569] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228670, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.541704] env[61768]: DEBUG nova.network.neutron [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 737.611420] env[61768]: DEBUG oslo_vmware.api [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228671, 'name': RemoveSnapshot_Task, 'duration_secs': 0.800846} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.611688] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 737.691246] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d911a93e-ba5c-4f4e-b119-957a180e260f tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.312s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.721860] env[61768]: DEBUG nova.network.neutron [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Updating instance_info_cache with network_info: [{"id": "79d4c508-cec4-40c8-8974-b1916ac2a4d2", "address": "fa:16:3e:b5:65:c7", "network": {"id": "09d8aff8-32f1-40c5-a201-5d949420729f", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-796992234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f446e9095b274a9abe29f764ff520164", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79d4c508-ce", "ovs_interfaceid": "79d4c508-cec4-40c8-8974-b1916ac2a4d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.925784] env[61768]: DEBUG nova.scheduler.client.report [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.932435] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228670, 'name': CreateVM_Task, 'duration_secs': 1.514518} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.932849] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 737.933581] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.933758] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.934141] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.934461] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-762cbf28-1585-473b-9362-ebbf93fda267 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.940939] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 737.940939] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529aea0f-ce2c-a7df-11c7-6e82c7aa099e" [ 737.940939] env[61768]: _type = "Task" [ 737.940939] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.951089] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529aea0f-ce2c-a7df-11c7-6e82c7aa099e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.039226] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 738.067686] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 738.067994] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 738.068141] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 738.068459] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 738.068623] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 738.068778] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 738.068997] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 738.069186] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 738.069395] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 738.069580] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 738.069759] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 738.071044] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aded7006-6985-49bd-891b-227dcfa5007d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.079711] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7bdff1-58ee-4bde-93c0-e7b899535823 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.117617] env[61768]: WARNING nova.compute.manager [None req-fbb24da0-79e2-4db2-b068-e6f53ff4f211 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Image not found during snapshot: nova.exception.ImageNotFound: Image 41d5ccba-d2fa-4a1f-8ebb-5cc3c3f1fb7e could not be found. [ 738.223054] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Releasing lock "refresh_cache-7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.223367] env[61768]: DEBUG nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Instance network_info: |[{"id": "79d4c508-cec4-40c8-8974-b1916ac2a4d2", "address": "fa:16:3e:b5:65:c7", "network": {"id": "09d8aff8-32f1-40c5-a201-5d949420729f", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-796992234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f446e9095b274a9abe29f764ff520164", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79d4c508-ce", "ovs_interfaceid": "79d4c508-cec4-40c8-8974-b1916ac2a4d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 738.223801] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:65:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '19671de9-8b5b-4710-adc3-7419f3c0f171', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '79d4c508-cec4-40c8-8974-b1916ac2a4d2', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 738.231532] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Creating folder: Project (f446e9095b274a9abe29f764ff520164). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 738.231940] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1768a04-a94f-417d-b6cb-8c005a9a49f3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.244563] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Created folder: Project (f446e9095b274a9abe29f764ff520164) in parent group-v265360. [ 738.244563] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Creating folder: Instances. Parent ref: group-v265403. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 738.244563] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0a55aca-45ad-4837-9074-7fab3e79ad8d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.253179] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Created folder: Instances in parent group-v265403. [ 738.253520] env[61768]: DEBUG oslo.service.loopingcall [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.253774] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 738.254009] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a3b75af3-0e30-4ee7-9cee-8ee0b3682b14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.274718] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 738.274718] env[61768]: value = "task-1228675" [ 738.274718] env[61768]: _type = "Task" [ 738.274718] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.283864] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228675, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.433910] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.434581] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.438447] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.117s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.439941] env[61768]: INFO nova.compute.claims [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.454791] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529aea0f-ce2c-a7df-11c7-6e82c7aa099e, 'name': SearchDatastore_Task, 'duration_secs': 0.013408} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.455172] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.455434] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 738.455709] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.455893] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.456124] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.456424] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-517cd24a-6741-4f47-bae5-ff2470d4868f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.461206] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.461464] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.461707] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.461932] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.462168] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.465377] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.465571] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 738.466688] env[61768]: INFO nova.compute.manager [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Terminating instance [ 738.468329] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30e5afde-4f93-4a86-aa1e-bb467165a9e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.471897] env[61768]: DEBUG nova.compute.manager [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.472151] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 738.472987] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6682d05d-aec7-40bf-bbf2-70bb6f4a3c6a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.480772] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 738.480772] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e7044e-0066-a915-f495-0fa19082e27e" [ 738.480772] env[61768]: _type = "Task" [ 738.480772] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.483465] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 738.486839] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec7a1c80-75e8-4536-ae05-bcf7121ea230 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.495331] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e7044e-0066-a915-f495-0fa19082e27e, 'name': SearchDatastore_Task, 'duration_secs': 0.011038} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.497430] env[61768]: DEBUG oslo_vmware.api [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 738.497430] env[61768]: value = "task-1228676" [ 738.497430] env[61768]: _type = "Task" [ 738.497430] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.498085] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9684b086-6375-4cba-9c69-51ebd554bc14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.506999] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 738.506999] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5233626b-190f-6f3e-38e4-3c999eee8ae2" [ 738.506999] env[61768]: _type = "Task" [ 738.506999] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.510670] env[61768]: DEBUG oslo_vmware.api [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228676, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.519713] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5233626b-190f-6f3e-38e4-3c999eee8ae2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.587575] env[61768]: DEBUG nova.compute.manager [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Received event network-changed-79d4c508-cec4-40c8-8974-b1916ac2a4d2 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 738.587789] env[61768]: DEBUG nova.compute.manager [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Refreshing instance network info cache due to event network-changed-79d4c508-cec4-40c8-8974-b1916ac2a4d2. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 738.587961] env[61768]: DEBUG oslo_concurrency.lockutils [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] Acquiring lock "refresh_cache-7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.588150] env[61768]: DEBUG oslo_concurrency.lockutils [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] Acquired lock "refresh_cache-7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.588338] env[61768]: DEBUG nova.network.neutron [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Refreshing network info cache for port 79d4c508-cec4-40c8-8974-b1916ac2a4d2 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.788023] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228675, 'name': CreateVM_Task, 'duration_secs': 0.431699} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.788023] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 738.788023] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.788023] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.788023] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 738.788023] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ffa26f2b-5b76-4570-b44b-aa1bfc7e6acd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.799041] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 738.799041] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c8e1ff-529f-d3d0-0881-94823fc5e943" [ 738.799041] env[61768]: _type = "Task" [ 738.799041] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.808855] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c8e1ff-529f-d3d0-0881-94823fc5e943, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.947542] env[61768]: DEBUG nova.compute.utils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.953558] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.953924] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 738.968832] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Successfully updated port: 3c8dc29a-e757-4d7d-93a4-9274629ec62c {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 739.011704] env[61768]: DEBUG oslo_vmware.api [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228676, 'name': PowerOffVM_Task, 'duration_secs': 0.22281} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.013323] env[61768]: DEBUG nova.policy [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1600e5a4cc4e44cb8f6d384d6fa1028a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54b7f349107b4f56960cae84edbea9e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 739.019211] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 739.019211] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 739.020639] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f76e5c3-0faa-479d-9d09-6f0fb1d45633 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.029139] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5233626b-190f-6f3e-38e4-3c999eee8ae2, 'name': SearchDatastore_Task, 'duration_secs': 0.016146} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.029406] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.029721] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 739.029946] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f85a236-d730-49a3-a95c-2f495da87e29 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.037362] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 739.037362] env[61768]: value = "task-1228678" [ 739.037362] env[61768]: _type = "Task" [ 739.037362] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.045186] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228678, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.095839] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 739.096903] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 739.097396] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Deleting the datastore file [datastore1] 91d56a8c-0664-4b3c-bf28-e2c668ac65fa {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.098141] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a56e6ab7-aa8e-4e16-9d69-6719fcd38269 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.106504] env[61768]: DEBUG oslo_vmware.api [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for the task: (returnval){ [ 739.106504] env[61768]: value = "task-1228679" [ 739.106504] env[61768]: _type = "Task" [ 739.106504] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.116096] env[61768]: DEBUG oslo_vmware.api [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228679, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.297633] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "0df161e9-4a66-432e-9366-60f231c53e3d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.297633] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "0df161e9-4a66-432e-9366-60f231c53e3d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.297633] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "0df161e9-4a66-432e-9366-60f231c53e3d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.297633] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "0df161e9-4a66-432e-9366-60f231c53e3d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.298186] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "0df161e9-4a66-432e-9366-60f231c53e3d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.299551] env[61768]: INFO nova.compute.manager [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Terminating instance [ 739.302585] env[61768]: DEBUG nova.compute.manager [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.303707] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 739.305176] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3ce858-a88f-4f53-92bb-995f09842fbb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.322394] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 739.327243] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-784e72e5-60b3-4e55-bb89-3d7e6281d074 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.329209] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c8e1ff-529f-d3d0-0881-94823fc5e943, 'name': SearchDatastore_Task, 'duration_secs': 0.010896} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.332395] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.332715] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 739.333153] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.333343] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.333571] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 739.334404] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fd3d897f-d412-42ef-b721-2c095b64511f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.342903] env[61768]: DEBUG oslo_vmware.api [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 739.342903] env[61768]: value = "task-1228680" [ 739.342903] env[61768]: _type = "Task" [ 739.342903] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.352295] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 739.352676] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 739.356064] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0c16268-00a1-4b6e-8797-4a57d061180e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.360267] env[61768]: DEBUG oslo_vmware.api [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228680, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.363729] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 739.363729] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5246387f-1e05-0e23-9980-262d142ce81d" [ 739.363729] env[61768]: _type = "Task" [ 739.363729] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.373069] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5246387f-1e05-0e23-9980-262d142ce81d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.435788] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Successfully created port: 16c13000-f458-4c8b-be59-ea8c7c28d9fc {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.455344] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.475569] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "refresh_cache-668d4fe4-85f8-4282-8d65-0549d9e3bda8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.475569] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "refresh_cache-668d4fe4-85f8-4282-8d65-0549d9e3bda8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.475835] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 739.486047] env[61768]: DEBUG nova.network.neutron [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Updated VIF entry in instance network info cache for port 79d4c508-cec4-40c8-8974-b1916ac2a4d2. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 739.486047] env[61768]: DEBUG nova.network.neutron [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Updating instance_info_cache with network_info: [{"id": "79d4c508-cec4-40c8-8974-b1916ac2a4d2", "address": "fa:16:3e:b5:65:c7", "network": {"id": "09d8aff8-32f1-40c5-a201-5d949420729f", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-796992234-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f446e9095b274a9abe29f764ff520164", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "19671de9-8b5b-4710-adc3-7419f3c0f171", "external-id": "nsx-vlan-transportzone-421", "segmentation_id": 421, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap79d4c508-ce", "ovs_interfaceid": "79d4c508-cec4-40c8-8974-b1916ac2a4d2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.547789] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228678, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.618620] env[61768]: DEBUG oslo_vmware.api [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Task: {'id': task-1228679, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.432444} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.618950] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 739.623028] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 739.623028] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 739.623028] env[61768]: INFO nova.compute.manager [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Took 1.15 seconds to destroy the instance on the hypervisor. [ 739.623028] env[61768]: DEBUG oslo.service.loopingcall [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.623028] env[61768]: DEBUG nova.compute.manager [-] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.623028] env[61768]: DEBUG nova.network.neutron [-] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.807474] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9fe2d61-99ca-4fc3-a62c-9eb7b4fd8382 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.816312] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb752b42-d54d-4b7b-ad37-22ec75f8d6f5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.853219] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-483182b1-4db8-4bbf-9f99-5a8ac1ff9658 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.866428] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7a1659-f221-4876-a267-2dc79680f2f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.871524] env[61768]: DEBUG oslo_vmware.api [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228680, 'name': PowerOffVM_Task, 'duration_secs': 0.20502} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.879140] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 739.879140] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 739.879140] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af80f945-ef59-42d3-9517-e98804222c0b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.888732] env[61768]: DEBUG nova.compute.provider_tree [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.894294] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5246387f-1e05-0e23-9980-262d142ce81d, 'name': SearchDatastore_Task, 'duration_secs': 0.052397} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.895337] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6dff62f9-8158-45ed-937a-e8c84919c524 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.901153] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 739.901153] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d3100-9ced-b7dc-26c7-db8ba96f9c9f" [ 739.901153] env[61768]: _type = "Task" [ 739.901153] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.909577] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d3100-9ced-b7dc-26c7-db8ba96f9c9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.953013] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 739.953970] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 739.953970] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleting the datastore file [datastore2] 0df161e9-4a66-432e-9366-60f231c53e3d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 739.953970] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c514de0b-0da1-4584-a15e-81859d1c27a0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.963614] env[61768]: DEBUG oslo_vmware.api [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 739.963614] env[61768]: value = "task-1228682" [ 739.963614] env[61768]: _type = "Task" [ 739.963614] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.984942] env[61768]: DEBUG oslo_vmware.api [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228682, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.985753] env[61768]: DEBUG oslo_concurrency.lockutils [req-c0952e74-61c0-407b-9961-d3e404cd60ef req-aae9234e-3d16-4c64-856a-6f4d62817a66 service nova] Releasing lock "refresh_cache-7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.050493] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228678, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515354} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.052021] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 740.052021] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 740.052021] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-62e378ab-8ef5-450f-8543-69cd3c97711e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.062315] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 740.062315] env[61768]: value = "task-1228683" [ 740.062315] env[61768]: _type = "Task" [ 740.062315] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.070973] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.188766] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 740.193024] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5214d399-a27b-f27e-a51f-fbf3d963bfe4/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 740.193723] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a8e7bbf-d135-489e-8f69-7a7178097a77 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.199932] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5214d399-a27b-f27e-a51f-fbf3d963bfe4/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 740.200134] env[61768]: ERROR oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5214d399-a27b-f27e-a51f-fbf3d963bfe4/disk-0.vmdk due to incomplete transfer. [ 740.200364] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-db5f4b79-61cb-4b05-a3a0-2711b9a3c77b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.207844] env[61768]: DEBUG oslo_vmware.rw_handles [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5214d399-a27b-f27e-a51f-fbf3d963bfe4/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 740.208184] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Uploaded image c5b4c304-600a-43d8-802c-259ae58cbf74 to the Glance image server {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 740.210366] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 740.212627] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-46bcc68d-a13f-4da1-acbc-59e97af69bb5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.218829] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 740.218829] env[61768]: value = "task-1228684" [ 740.218829] env[61768]: _type = "Task" [ 740.218829] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.227106] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228684, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.378648] env[61768]: DEBUG nova.network.neutron [-] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.393743] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Updating instance_info_cache with network_info: [{"id": "3c8dc29a-e757-4d7d-93a4-9274629ec62c", "address": "fa:16:3e:d1:93:05", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8dc29a-e7", "ovs_interfaceid": "3c8dc29a-e757-4d7d-93a4-9274629ec62c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.396353] env[61768]: DEBUG nova.scheduler.client.report [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.418470] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d3100-9ced-b7dc-26c7-db8ba96f9c9f, 'name': SearchDatastore_Task, 'duration_secs': 0.011492} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.418752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.419077] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3/7ef50ee4-64e6-49c2-bb41-0bdce316f8d3.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 740.419568] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-712dd090-2494-4d45-a5de-7188c99d2ee8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.429133] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 740.429133] env[61768]: value = "task-1228685" [ 740.429133] env[61768]: _type = "Task" [ 740.429133] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.438207] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228685, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.465660] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.479309] env[61768]: DEBUG oslo_vmware.api [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228682, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191638} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.480555] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.480555] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 740.480555] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 740.480555] env[61768]: INFO nova.compute.manager [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Took 1.18 seconds to destroy the instance on the hypervisor. [ 740.480935] env[61768]: DEBUG oslo.service.loopingcall [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.480935] env[61768]: DEBUG nova.compute.manager [-] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 740.481039] env[61768]: DEBUG nova.network.neutron [-] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 740.497484] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.497759] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.497925] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.498129] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.498542] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.498542] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.498670] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.498811] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.498996] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.499197] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.499421] env[61768]: DEBUG nova.virt.hardware [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.500315] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9e22fd2-3481-497d-8773-63aa0865f70a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.508587] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d495138-b6ad-4d8b-8461-f35cb1c24132 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.571354] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074826} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.571716] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 740.572555] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2f8c25-3143-4204-a7eb-4fd3e190d1a7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.599056] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 740.600669] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a94f63eb-e71b-4d1c-bff6-b90ddb8c3456 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.626814] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 740.626814] env[61768]: value = "task-1228686" [ 740.626814] env[61768]: _type = "Task" [ 740.626814] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.638033] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228686, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.729504] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228684, 'name': Destroy_Task, 'duration_secs': 0.387277} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.729810] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Destroyed the VM [ 740.730065] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 740.730332] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4af9058b-cab7-4b8f-b2df-43d5dd29fbe4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.737892] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 740.737892] env[61768]: value = "task-1228687" [ 740.737892] env[61768]: _type = "Task" [ 740.737892] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.746753] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228687, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.749409] env[61768]: DEBUG nova.compute.manager [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Received event network-vif-plugged-3c8dc29a-e757-4d7d-93a4-9274629ec62c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.749657] env[61768]: DEBUG oslo_concurrency.lockutils [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] Acquiring lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.749874] env[61768]: DEBUG oslo_concurrency.lockutils [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.750061] env[61768]: DEBUG oslo_concurrency.lockutils [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.750301] env[61768]: DEBUG nova.compute.manager [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] No waiting events found dispatching network-vif-plugged-3c8dc29a-e757-4d7d-93a4-9274629ec62c {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 740.750575] env[61768]: WARNING nova.compute.manager [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Received unexpected event network-vif-plugged-3c8dc29a-e757-4d7d-93a4-9274629ec62c for instance with vm_state building and task_state spawning. [ 740.750633] env[61768]: DEBUG nova.compute.manager [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Received event network-changed-3c8dc29a-e757-4d7d-93a4-9274629ec62c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 740.750807] env[61768]: DEBUG nova.compute.manager [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Refreshing instance network info cache due to event network-changed-3c8dc29a-e757-4d7d-93a4-9274629ec62c. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 740.750957] env[61768]: DEBUG oslo_concurrency.lockutils [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] Acquiring lock "refresh_cache-668d4fe4-85f8-4282-8d65-0549d9e3bda8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.883716] env[61768]: INFO nova.compute.manager [-] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Took 1.26 seconds to deallocate network for instance. [ 740.905036] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "refresh_cache-668d4fe4-85f8-4282-8d65-0549d9e3bda8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.905036] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Instance network_info: |[{"id": "3c8dc29a-e757-4d7d-93a4-9274629ec62c", "address": "fa:16:3e:d1:93:05", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8dc29a-e7", "ovs_interfaceid": "3c8dc29a-e757-4d7d-93a4-9274629ec62c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 740.905317] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.465s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.905317] env[61768]: DEBUG nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.906536] env[61768]: DEBUG oslo_concurrency.lockutils [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] Acquired lock "refresh_cache-668d4fe4-85f8-4282-8d65-0549d9e3bda8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.907353] env[61768]: DEBUG nova.network.neutron [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Refreshing network info cache for port 3c8dc29a-e757-4d7d-93a4-9274629ec62c {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.909635] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:93:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31a7f15-a808-4199-9071-31fd05e316ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c8dc29a-e757-4d7d-93a4-9274629ec62c', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 740.919110] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Creating folder: Project (54b7f349107b4f56960cae84edbea9e0). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 740.919802] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.592s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.921511] env[61768]: INFO nova.compute.claims [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.928566] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-740f8831-89cc-4889-a35e-25e44d46cc80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.948253] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228685, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.950198] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Created folder: Project (54b7f349107b4f56960cae84edbea9e0) in parent group-v265360. [ 740.950615] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Creating folder: Instances. Parent ref: group-v265406. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 740.951214] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-27d138cf-1f77-429e-ac0d-78be70fd5af1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.965049] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Created folder: Instances in parent group-v265406. [ 740.965049] env[61768]: DEBUG oslo.service.loopingcall [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 740.965049] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 740.965049] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e41bbef-cb39-4ecc-bc5c-25f6d12bf24a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.988945] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 740.988945] env[61768]: value = "task-1228690" [ 740.988945] env[61768]: _type = "Task" [ 740.988945] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.999597] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228690, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.123859] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.123859] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.123859] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.123859] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.124299] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.124716] env[61768]: INFO nova.compute.manager [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Terminating instance [ 741.127490] env[61768]: DEBUG nova.compute.manager [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 741.127490] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 741.128444] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dff8630-c835-4043-bb0f-329d17e1b7da {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.142302] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228686, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.144882] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 741.145723] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2a59992-fc06-498d-a182-7fc1fd64a584 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.191586] env[61768]: DEBUG nova.compute.manager [req-955a8a65-ef33-421a-8cd2-8d5c0e3940a4 req-1d8ba8ee-db15-40d1-a554-9369bdd7f2a7 service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Received event network-vif-plugged-16c13000-f458-4c8b-be59-ea8c7c28d9fc {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.191908] env[61768]: DEBUG oslo_concurrency.lockutils [req-955a8a65-ef33-421a-8cd2-8d5c0e3940a4 req-1d8ba8ee-db15-40d1-a554-9369bdd7f2a7 service nova] Acquiring lock "9fa32a4d-3ba4-4d36-963b-17a64453e804-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.192266] env[61768]: DEBUG oslo_concurrency.lockutils [req-955a8a65-ef33-421a-8cd2-8d5c0e3940a4 req-1d8ba8ee-db15-40d1-a554-9369bdd7f2a7 service nova] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.192540] env[61768]: DEBUG oslo_concurrency.lockutils [req-955a8a65-ef33-421a-8cd2-8d5c0e3940a4 req-1d8ba8ee-db15-40d1-a554-9369bdd7f2a7 service nova] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.192804] env[61768]: DEBUG nova.compute.manager [req-955a8a65-ef33-421a-8cd2-8d5c0e3940a4 req-1d8ba8ee-db15-40d1-a554-9369bdd7f2a7 service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] No waiting events found dispatching network-vif-plugged-16c13000-f458-4c8b-be59-ea8c7c28d9fc {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 741.193093] env[61768]: WARNING nova.compute.manager [req-955a8a65-ef33-421a-8cd2-8d5c0e3940a4 req-1d8ba8ee-db15-40d1-a554-9369bdd7f2a7 service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Received unexpected event network-vif-plugged-16c13000-f458-4c8b-be59-ea8c7c28d9fc for instance with vm_state building and task_state spawning. [ 741.218265] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Successfully updated port: 16c13000-f458-4c8b-be59-ea8c7c28d9fc {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 741.252194] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228687, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.324078] env[61768]: DEBUG nova.network.neutron [-] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.390782] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.433729] env[61768]: DEBUG nova.compute.utils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.435360] env[61768]: DEBUG nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.435564] env[61768]: DEBUG nova.network.neutron [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 741.447210] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228685, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.568708} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.447482] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3/7ef50ee4-64e6-49c2-bb41-0bdce316f8d3.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 741.447710] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 741.447949] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ccf95756-e072-4769-94dd-b1f6dfe5d8cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.454452] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 741.454452] env[61768]: value = "task-1228692" [ 741.454452] env[61768]: _type = "Task" [ 741.454452] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.466861] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228692, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.485094] env[61768]: DEBUG nova.policy [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '471950efcaee4db0b13f213bb01c770d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17cebbaa8aa846b1b7fafce4782b2622', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.499817] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228690, 'name': CreateVM_Task, 'duration_secs': 0.425073} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.500031] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 741.500850] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.501106] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.501492] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 741.501769] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-371867bd-bd0b-4de0-873e-4678d8b202d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.509152] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 741.509152] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5226f56b-0dda-0790-a07b-2d079fed2d8a" [ 741.509152] env[61768]: _type = "Task" [ 741.509152] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.519421] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5226f56b-0dda-0790-a07b-2d079fed2d8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.594088] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 741.594300] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 741.594492] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleting the datastore file [datastore2] 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 741.594768] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92330237-cccf-47c0-b234-aaa451bc5db6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.601195] env[61768]: DEBUG oslo_vmware.api [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 741.601195] env[61768]: value = "task-1228693" [ 741.601195] env[61768]: _type = "Task" [ 741.601195] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.358039] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "refresh_cache-9fa32a4d-3ba4-4d36-963b-17a64453e804" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.358255] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "refresh_cache-9fa32a4d-3ba4-4d36-963b-17a64453e804" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.358390] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.359635] env[61768]: INFO nova.compute.manager [-] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Took 1.88 seconds to deallocate network for instance. [ 742.362518] env[61768]: DEBUG nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.365272] env[61768]: DEBUG oslo_vmware.api [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228693, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.365460] env[61768]: WARNING oslo_vmware.common.loopingcall [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] task run outlasted interval by 0.26404099999999997 sec [ 742.368608] env[61768]: DEBUG nova.network.neutron [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Successfully created port: be7be29f-b1a1-4851-b1a4-b368519c25ef {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.388579] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.388821] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.406251] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228692, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076222} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.406754] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228686, 'name': ReconfigVM_Task, 'duration_secs': 0.69719} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.411632] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 742.412083] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Reconfigured VM instance instance-00000032 to attach disk [datastore1] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 742.413425] env[61768]: DEBUG oslo_vmware.api [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228687, 'name': RemoveSnapshot_Task, 'duration_secs': 0.714335} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.416863] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ce6ec92-927e-4764-8a1a-edaba7954afa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.419386] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52221ff1-a46d-4912-90a4-b1b796db7ffc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.421249] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 742.421498] env[61768]: INFO nova.compute.manager [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Took 16.07 seconds to snapshot the instance on the hypervisor. [ 742.424433] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5226f56b-0dda-0790-a07b-2d079fed2d8a, 'name': SearchDatastore_Task, 'duration_secs': 0.02598} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.431612] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.431612] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.431772] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.431881] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.432075] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.433016] env[61768]: DEBUG oslo_vmware.api [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228693, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152714} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.433942] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8edf2518-d3be-4fb5-ba89-aff2abe04587 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.439018] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 742.439018] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 742.439018] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 742.439018] env[61768]: INFO nova.compute.manager [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Took 1.31 seconds to destroy the instance on the hypervisor. [ 742.439018] env[61768]: DEBUG oslo.service.loopingcall [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 742.446151] env[61768]: DEBUG nova.compute.manager [-] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 742.446268] env[61768]: DEBUG nova.network.neutron [-] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.456136] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3/7ef50ee4-64e6-49c2-bb41-0bdce316f8d3.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 742.460990] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70beadf6-b7bc-4de9-a777-c50c4832e4c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.476207] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 742.476207] env[61768]: value = "task-1228694" [ 742.476207] env[61768]: _type = "Task" [ 742.476207] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.480417] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.480633] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 742.485505] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-111bafc5-66d1-43ef-aa96-a4cf25ec023b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.489465] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 742.489465] env[61768]: value = "task-1228695" [ 742.489465] env[61768]: _type = "Task" [ 742.489465] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.500283] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 742.500283] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52dbc715-0eb1-eaf0-6109-2e14683f5198" [ 742.500283] env[61768]: _type = "Task" [ 742.500283] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.500509] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228694, 'name': Rename_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.507624] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228695, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.513023] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52dbc715-0eb1-eaf0-6109-2e14683f5198, 'name': SearchDatastore_Task, 'duration_secs': 0.012675} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.516779] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f10db76a-7265-413f-9e02-eddeb22aeb00 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.521594] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 742.521594] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522a9c47-b6a4-18a0-664a-3d65f76e0784" [ 742.521594] env[61768]: _type = "Task" [ 742.521594] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.528671] env[61768]: DEBUG nova.network.neutron [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Updated VIF entry in instance network info cache for port 3c8dc29a-e757-4d7d-93a4-9274629ec62c. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 742.528671] env[61768]: DEBUG nova.network.neutron [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Updating instance_info_cache with network_info: [{"id": "3c8dc29a-e757-4d7d-93a4-9274629ec62c", "address": "fa:16:3e:d1:93:05", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8dc29a-e7", "ovs_interfaceid": "3c8dc29a-e757-4d7d-93a4-9274629ec62c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.536573] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522a9c47-b6a4-18a0-664a-3d65f76e0784, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.760556] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e2abea-d88b-4e9c-8afa-7b1975cd1146 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.768637] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ba6216-b91d-4c67-a731-117374d2687b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.799655] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13788046-79a8-4348-9234-42248b2b5d27 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.803340] env[61768]: DEBUG nova.compute.manager [req-5949f714-be01-481b-b2ea-2cdeaf88c74e req-62996279-5498-4ebf-84dd-ec9908a491b9 service nova] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Received event network-vif-deleted-4b6df580-88b9-4455-b18b-0a48b05e7f73 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.809585] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76cfefc3-16fd-44ed-b06a-995a45119ab7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.823573] env[61768]: DEBUG nova.compute.provider_tree [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.895997] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.918359] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.928612] env[61768]: DEBUG nova.compute.manager [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Instance disappeared during snapshot {{(pid=61768) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 742.945115] env[61768]: DEBUG nova.compute.manager [None req-69eeb384-9124-457e-a7bd-a8c789422926 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image not found during clean up c5b4c304-600a-43d8-802c-259ae58cbf74 {{(pid=61768) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 742.989718] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228694, 'name': Rename_Task, 'duration_secs': 0.284457} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.992196] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 742.992553] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d88298d-8ce6-40a8-894a-c49bbdcdf839 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.002260] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228695, 'name': ReconfigVM_Task, 'duration_secs': 0.49275} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.003433] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3/7ef50ee4-64e6-49c2-bb41-0bdce316f8d3.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 743.004127] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 743.004127] env[61768]: value = "task-1228696" [ 743.004127] env[61768]: _type = "Task" [ 743.004127] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.004323] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-498ad450-6ddc-4d8b-b747-edf57609f300 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.013483] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228696, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.014664] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 743.014664] env[61768]: value = "task-1228697" [ 743.014664] env[61768]: _type = "Task" [ 743.014664] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.022965] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228697, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.031119] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522a9c47-b6a4-18a0-664a-3d65f76e0784, 'name': SearchDatastore_Task, 'duration_secs': 0.014398} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.031392] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.031657] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 668d4fe4-85f8-4282-8d65-0549d9e3bda8/668d4fe4-85f8-4282-8d65-0549d9e3bda8.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 743.032148] env[61768]: DEBUG oslo_concurrency.lockutils [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] Releasing lock "refresh_cache-668d4fe4-85f8-4282-8d65-0549d9e3bda8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.032402] env[61768]: DEBUG nova.compute.manager [req-7e1f71b5-b753-451e-9675-4a4177eb025c req-ddf961f6-359c-49ad-bf9e-eb7d8ac8833c service nova] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Received event network-vif-deleted-932b7671-16df-463c-b31c-e99ede6a0287 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.032752] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-80d278e8-3f0e-4985-bd14-97ee6ad7bf07 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.038916] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 743.038916] env[61768]: value = "task-1228698" [ 743.038916] env[61768]: _type = "Task" [ 743.038916] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.047363] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228698, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.164884] env[61768]: DEBUG nova.network.neutron [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Updating instance_info_cache with network_info: [{"id": "16c13000-f458-4c8b-be59-ea8c7c28d9fc", "address": "fa:16:3e:95:d0:c9", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16c13000-f4", "ovs_interfaceid": "16c13000-f458-4c8b-be59-ea8c7c28d9fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.266351] env[61768]: DEBUG nova.compute.manager [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Received event network-changed-16c13000-f458-4c8b-be59-ea8c7c28d9fc {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.266571] env[61768]: DEBUG nova.compute.manager [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Refreshing instance network info cache due to event network-changed-16c13000-f458-4c8b-be59-ea8c7c28d9fc. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 743.266778] env[61768]: DEBUG oslo_concurrency.lockutils [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] Acquiring lock "refresh_cache-9fa32a4d-3ba4-4d36-963b-17a64453e804" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.327301] env[61768]: DEBUG nova.scheduler.client.report [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 743.342287] env[61768]: DEBUG nova.network.neutron [-] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.378416] env[61768]: DEBUG nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.408483] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.408811] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.409084] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.409178] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.409374] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.409597] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.409905] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.410122] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.410321] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.410498] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.410680] env[61768]: DEBUG nova.virt.hardware [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.411657] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba52236-fd8b-4772-80fd-5f96509bba12 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.421670] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85d67102-0df5-4b79-8672-964d4e401564 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.516348] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228696, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.525403] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228697, 'name': Rename_Task, 'duration_secs': 0.14006} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.526146] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 743.526439] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bcaad6f7-6473-47cf-8cc6-f45659b69108 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.540139] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 743.540139] env[61768]: value = "task-1228699" [ 743.540139] env[61768]: _type = "Task" [ 743.540139] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.553536] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228699, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.556697] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228698, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.670093] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "refresh_cache-9fa32a4d-3ba4-4d36-963b-17a64453e804" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.670432] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Instance network_info: |[{"id": "16c13000-f458-4c8b-be59-ea8c7c28d9fc", "address": "fa:16:3e:95:d0:c9", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16c13000-f4", "ovs_interfaceid": "16c13000-f458-4c8b-be59-ea8c7c28d9fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 743.670760] env[61768]: DEBUG oslo_concurrency.lockutils [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] Acquired lock "refresh_cache-9fa32a4d-3ba4-4d36-963b-17a64453e804" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.670948] env[61768]: DEBUG nova.network.neutron [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Refreshing network info cache for port 16c13000-f458-4c8b-be59-ea8c7c28d9fc {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 743.672734] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:d0:c9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31a7f15-a808-4199-9071-31fd05e316ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '16c13000-f458-4c8b-be59-ea8c7c28d9fc', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.679904] env[61768]: DEBUG oslo.service.loopingcall [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.680375] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 743.680601] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8241872e-33e8-4871-9a4c-da2c7800754d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.701391] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.701391] env[61768]: value = "task-1228700" [ 743.701391] env[61768]: _type = "Task" [ 743.701391] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.709555] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228700, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.833925] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.914s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.834684] env[61768]: DEBUG nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.837723] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.971s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.839229] env[61768]: INFO nova.compute.claims [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.844082] env[61768]: INFO nova.compute.manager [-] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Took 1.40 seconds to deallocate network for instance. [ 744.016918] env[61768]: DEBUG oslo_vmware.api [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228696, 'name': PowerOnVM_Task, 'duration_secs': 0.742485} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.017252] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 744.017471] env[61768]: INFO nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Took 11.01 seconds to spawn the instance on the hypervisor. [ 744.017635] env[61768]: DEBUG nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 744.018518] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc3bfec-380a-40b2-ab5f-dd378a5bdfeb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.052486] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228699, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.055769] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228698, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537488} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.056042] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 668d4fe4-85f8-4282-8d65-0549d9e3bda8/668d4fe4-85f8-4282-8d65-0549d9e3bda8.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 744.056272] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.056521] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b6275c59-bb62-416d-b451-6816bf8f81b8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.063825] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 744.063825] env[61768]: value = "task-1228701" [ 744.063825] env[61768]: _type = "Task" [ 744.063825] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.071681] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228701, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.102148] env[61768]: DEBUG nova.network.neutron [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Successfully updated port: be7be29f-b1a1-4851-b1a4-b368519c25ef {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 744.214902] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228700, 'name': CreateVM_Task, 'duration_secs': 0.502716} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.217198] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 744.217914] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.218143] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.218480] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 744.218744] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8b50ef31-42f7-4f21-b3cc-247e7343b094 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.224057] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 744.224057] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5283fc9b-f929-3cbe-fab9-d539cf433150" [ 744.224057] env[61768]: _type = "Task" [ 744.224057] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.232208] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5283fc9b-f929-3cbe-fab9-d539cf433150, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.343897] env[61768]: DEBUG nova.compute.utils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 744.347708] env[61768]: DEBUG nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 744.347876] env[61768]: DEBUG nova.network.neutron [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 744.352109] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.439675] env[61768]: DEBUG nova.policy [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf57ec11d6374eab992984853f85e87b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ec5c061ca09040e8a2b3c3a6874daa40', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 744.534761] env[61768]: INFO nova.compute.manager [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Took 36.71 seconds to build instance. [ 744.550023] env[61768]: DEBUG nova.network.neutron [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Updated VIF entry in instance network info cache for port 16c13000-f458-4c8b-be59-ea8c7c28d9fc. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 744.551112] env[61768]: DEBUG nova.network.neutron [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Updating instance_info_cache with network_info: [{"id": "16c13000-f458-4c8b-be59-ea8c7c28d9fc", "address": "fa:16:3e:95:d0:c9", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap16c13000-f4", "ovs_interfaceid": "16c13000-f458-4c8b-be59-ea8c7c28d9fc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.557627] env[61768]: DEBUG oslo_vmware.api [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228699, 'name': PowerOnVM_Task, 'duration_secs': 0.768826} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.558222] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 744.558539] env[61768]: INFO nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Took 9.05 seconds to spawn the instance on the hypervisor. [ 744.558940] env[61768]: DEBUG nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 744.560355] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8052bf6-3d53-4669-bd0c-49ec2a0b1c4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.578936] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228701, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066708} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.579542] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 744.581130] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a35d624-6064-4953-a8e1-cec60b884ace {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.604631] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] 668d4fe4-85f8-4282-8d65-0549d9e3bda8/668d4fe4-85f8-4282-8d65-0549d9e3bda8.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 744.606022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "refresh_cache-9b3632e5-4c2d-4968-9692-70f91883bfb3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.606022] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquired lock "refresh_cache-9b3632e5-4c2d-4968-9692-70f91883bfb3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.606022] env[61768]: DEBUG nova.network.neutron [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 744.607658] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-386cb2dc-ad92-4e3d-b1ac-0434022f11a0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.632018] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 744.632018] env[61768]: value = "task-1228702" [ 744.632018] env[61768]: _type = "Task" [ 744.632018] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.639930] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228702, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.734850] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5283fc9b-f929-3cbe-fab9-d539cf433150, 'name': SearchDatastore_Task, 'duration_secs': 0.030075} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.735251] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.735496] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.735738] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.736291] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.736291] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.736654] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95e9f22d-f747-4467-a161-e5b12137b148 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.750062] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.750304] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 744.751037] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a26747c-5f8d-40e8-9519-22991eed069b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.757118] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 744.757118] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52362897-57f6-c98a-0918-32a04568465f" [ 744.757118] env[61768]: _type = "Task" [ 744.757118] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.764518] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52362897-57f6-c98a-0918-32a04568465f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.771930] env[61768]: DEBUG nova.network.neutron [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Successfully created port: 61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.817024] env[61768]: DEBUG nova.compute.manager [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Received event network-vif-deleted-59dd0ec8-7838-4fa2-ac70-42482a388241 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.817322] env[61768]: DEBUG nova.compute.manager [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Received event network-vif-plugged-be7be29f-b1a1-4851-b1a4-b368519c25ef {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.817522] env[61768]: DEBUG oslo_concurrency.lockutils [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] Acquiring lock "9b3632e5-4c2d-4968-9692-70f91883bfb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.817762] env[61768]: DEBUG oslo_concurrency.lockutils [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.817904] env[61768]: DEBUG oslo_concurrency.lockutils [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.818116] env[61768]: DEBUG nova.compute.manager [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] No waiting events found dispatching network-vif-plugged-be7be29f-b1a1-4851-b1a4-b368519c25ef {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 744.818440] env[61768]: WARNING nova.compute.manager [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Received unexpected event network-vif-plugged-be7be29f-b1a1-4851-b1a4-b368519c25ef for instance with vm_state building and task_state spawning. [ 744.818670] env[61768]: DEBUG nova.compute.manager [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Received event network-changed-be7be29f-b1a1-4851-b1a4-b368519c25ef {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.818789] env[61768]: DEBUG nova.compute.manager [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Refreshing instance network info cache due to event network-changed-be7be29f-b1a1-4851-b1a4-b368519c25ef. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 744.819059] env[61768]: DEBUG oslo_concurrency.lockutils [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] Acquiring lock "refresh_cache-9b3632e5-4c2d-4968-9692-70f91883bfb3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.848224] env[61768]: DEBUG nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 745.037627] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94de5d8c-4b82-4116-bb5c-b54d722c5a93 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "efbde802-8cb0-4563-a776-3722a3889afe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 142.201s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.058947] env[61768]: DEBUG oslo_concurrency.lockutils [req-f09362fa-db52-4b91-90fc-5d3f542647f8 req-553f2731-6d00-4bc8-be03-99c1a5e01ccb service nova] Releasing lock "refresh_cache-9fa32a4d-3ba4-4d36-963b-17a64453e804" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.082100] env[61768]: INFO nova.compute.manager [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Took 35.33 seconds to build instance. [ 745.136827] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da65d91b-0ae5-4990-8cd9-39bbc78c3400 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.139875] env[61768]: DEBUG nova.network.neutron [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.144414] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228702, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.148728] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-228b2232-3961-4a2e-b43f-cb2307420e39 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.182588] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ebeba38-f948-4886-9631-93d3333b040c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.190208] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca4e11fc-5ac6-4632-bf8c-01830ca671ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.203505] env[61768]: DEBUG nova.compute.provider_tree [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.269976] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52362897-57f6-c98a-0918-32a04568465f, 'name': SearchDatastore_Task, 'duration_secs': 0.025102} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.270797] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d7a2d67-146c-4ef2-b4ba-fff74b013eef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.276177] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 745.276177] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a59943-9fec-adee-2257-c36b1b0fca7d" [ 745.276177] env[61768]: _type = "Task" [ 745.276177] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.284881] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a59943-9fec-adee-2257-c36b1b0fca7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.350889] env[61768]: DEBUG nova.network.neutron [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Updating instance_info_cache with network_info: [{"id": "be7be29f-b1a1-4851-b1a4-b368519c25ef", "address": "fa:16:3e:0d:f0:c4", "network": {"id": "697af0e3-d4e4-4005-bfba-9520210b422b", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1580881405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17cebbaa8aa846b1b7fafce4782b2622", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe7be29f-b1", "ovs_interfaceid": "be7be29f-b1a1-4851-b1a4-b368519c25ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.540591] env[61768]: DEBUG nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 745.584424] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bcfdf2e9-6233-4225-b597-c8176aa2147c tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.636s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.640939] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228702, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.706699] env[61768]: DEBUG nova.scheduler.client.report [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 745.787709] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a59943-9fec-adee-2257-c36b1b0fca7d, 'name': SearchDatastore_Task, 'duration_secs': 0.011918} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.787996] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.788325] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 9fa32a4d-3ba4-4d36-963b-17a64453e804/9fa32a4d-3ba4-4d36-963b-17a64453e804.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 745.788629] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e95250a8-6461-4e09-8b61-2e952885f4e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.796263] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 745.796263] env[61768]: value = "task-1228703" [ 745.796263] env[61768]: _type = "Task" [ 745.796263] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.805415] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.856664] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Releasing lock "refresh_cache-9b3632e5-4c2d-4968-9692-70f91883bfb3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.856664] env[61768]: DEBUG nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Instance network_info: |[{"id": "be7be29f-b1a1-4851-b1a4-b368519c25ef", "address": "fa:16:3e:0d:f0:c4", "network": {"id": "697af0e3-d4e4-4005-bfba-9520210b422b", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1580881405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17cebbaa8aa846b1b7fafce4782b2622", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe7be29f-b1", "ovs_interfaceid": "be7be29f-b1a1-4851-b1a4-b368519c25ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 745.856787] env[61768]: DEBUG oslo_concurrency.lockutils [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] Acquired lock "refresh_cache-9b3632e5-4c2d-4968-9692-70f91883bfb3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.856787] env[61768]: DEBUG nova.network.neutron [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Refreshing network info cache for port be7be29f-b1a1-4851-b1a4-b368519c25ef {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 745.856787] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:f0:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '650f7968-4522-4ba5-8304-1b9949951ed7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be7be29f-b1a1-4851-b1a4-b368519c25ef', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 745.864023] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Creating folder: Project (17cebbaa8aa846b1b7fafce4782b2622). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 745.869481] env[61768]: DEBUG nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.872094] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7da5ac50-2a92-44ea-8f3f-df71a4ac0746 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.883937] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Created folder: Project (17cebbaa8aa846b1b7fafce4782b2622) in parent group-v265360. [ 745.884231] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Creating folder: Instances. Parent ref: group-v265410. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 745.884603] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9627d775-4f0c-4165-a1e3-e667b2d0da6d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.893924] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Created folder: Instances in parent group-v265410. [ 745.894216] env[61768]: DEBUG oslo.service.loopingcall [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.894522] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 745.894748] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-628f5eab-3406-4a34-89c8-427af88b816b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.918035] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 745.918035] env[61768]: value = "task-1228706" [ 745.918035] env[61768]: _type = "Task" [ 745.918035] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.924091] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.924247] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.924369] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.924557] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.924710] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.925024] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.925248] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.925417] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.925591] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.925758] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.925933] env[61768]: DEBUG nova.virt.hardware [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.926767] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c99c1bb7-bee1-4b88-bedd-83a5944cb563 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.934370] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228706, 'name': CreateVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.941206] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1366cb31-702a-41ff-9b76-3b0b09eba9b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.067577] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.088303] env[61768]: DEBUG nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 746.143153] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228702, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.212239] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.212544] env[61768]: DEBUG nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 746.216822] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.903s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.218747] env[61768]: INFO nova.compute.claims [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.300600] env[61768]: DEBUG nova.network.neutron [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Updated VIF entry in instance network info cache for port be7be29f-b1a1-4851-b1a4-b368519c25ef. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 746.301851] env[61768]: DEBUG nova.network.neutron [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Updating instance_info_cache with network_info: [{"id": "be7be29f-b1a1-4851-b1a4-b368519c25ef", "address": "fa:16:3e:0d:f0:c4", "network": {"id": "697af0e3-d4e4-4005-bfba-9520210b422b", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1580881405-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17cebbaa8aa846b1b7fafce4782b2622", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "650f7968-4522-4ba5-8304-1b9949951ed7", "external-id": "nsx-vlan-transportzone-568", "segmentation_id": 568, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe7be29f-b1", "ovs_interfaceid": "be7be29f-b1a1-4851-b1a4-b368519c25ef", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.312295] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228703, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.429863] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228706, 'name': CreateVM_Task, 'duration_secs': 0.381164} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.429948] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 746.430747] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.430922] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.431338] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 746.431558] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea8bcafa-eb1a-4317-a354-7c53599cf23d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.438534] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 746.438534] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5293d8a6-94e6-ff1c-ce6a-9c1b6c87ab4d" [ 746.438534] env[61768]: _type = "Task" [ 746.438534] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.450922] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5293d8a6-94e6-ff1c-ce6a-9c1b6c87ab4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.610955] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.644956] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228702, 'name': ReconfigVM_Task, 'duration_secs': 1.710461} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.644956] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Reconfigured VM instance instance-00000034 to attach disk [datastore1] 668d4fe4-85f8-4282-8d65-0549d9e3bda8/668d4fe4-85f8-4282-8d65-0549d9e3bda8.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.644956] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95cbf1eb-3394-4dff-847a-1352bf644d9d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.653705] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 746.653705] env[61768]: value = "task-1228707" [ 746.653705] env[61768]: _type = "Task" [ 746.653705] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.664883] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228707, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.676659] env[61768]: DEBUG nova.network.neutron [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Successfully updated port: 61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 746.725899] env[61768]: DEBUG nova.compute.utils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 746.729850] env[61768]: DEBUG nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 746.730063] env[61768]: DEBUG nova.network.neutron [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 746.802540] env[61768]: DEBUG nova.policy [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4b31a829aea14248baf2df0c663eae30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '272b4bfd904e4bfebb692b0492f955c8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 746.806661] env[61768]: DEBUG oslo_concurrency.lockutils [req-76bc2b20-69b1-4785-9fc4-2ba9793c30db req-8a8fbc49-1e89-4f14-a730-627652843c0d service nova] Releasing lock "refresh_cache-9b3632e5-4c2d-4968-9692-70f91883bfb3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.811283] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228703, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.913447} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.811531] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 9fa32a4d-3ba4-4d36-963b-17a64453e804/9fa32a4d-3ba4-4d36-963b-17a64453e804.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 746.811745] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 746.811995] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59a02611-925e-4513-9a55-fceff530af8e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.819435] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 746.819435] env[61768]: value = "task-1228708" [ 746.819435] env[61768]: _type = "Task" [ 746.819435] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.834392] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228708, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.909837] env[61768]: INFO nova.compute.manager [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Rebuilding instance [ 746.949671] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5293d8a6-94e6-ff1c-ce6a-9c1b6c87ab4d, 'name': SearchDatastore_Task, 'duration_secs': 0.052545} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.956411] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.956740] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 746.957010] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.957201] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.957407] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 746.957902] env[61768]: DEBUG nova.compute.manager [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 746.958172] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4d46721-6334-42b9-8e41-abd1df637c1d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.961207] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27707c5-703b-4b7a-ad2f-987233e3180c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.970531] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 746.970676] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 746.972118] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a6a2e4e-32a1-437c-a5aa-f487f25524f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.977917] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 746.977917] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fbf610-ee24-c9ca-dd8b-8dde7ebe9a74" [ 746.977917] env[61768]: _type = "Task" [ 746.977917] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.986054] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fbf610-ee24-c9ca-dd8b-8dde7ebe9a74, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.092527] env[61768]: DEBUG nova.compute.manager [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received event network-vif-plugged-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.092819] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] Acquiring lock "23a4450d-162a-4b2b-a009-7023851315a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.092991] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] Lock "23a4450d-162a-4b2b-a009-7023851315a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.093193] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] Lock "23a4450d-162a-4b2b-a009-7023851315a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.093372] env[61768]: DEBUG nova.compute.manager [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] No waiting events found dispatching network-vif-plugged-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 747.093544] env[61768]: WARNING nova.compute.manager [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received unexpected event network-vif-plugged-61e8c28f-c389-4d6d-af8e-dabbc4c83363 for instance with vm_state building and task_state spawning. [ 747.093711] env[61768]: DEBUG nova.compute.manager [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.093939] env[61768]: DEBUG nova.compute.manager [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing instance network info cache due to event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 747.094164] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] Acquiring lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.094310] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] Acquired lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.094473] env[61768]: DEBUG nova.network.neutron [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 747.167693] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228707, 'name': Rename_Task, 'duration_secs': 0.160248} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.167777] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 747.167995] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-354a17db-2e65-43d9-bad8-362b36f2319e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.174250] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 747.174250] env[61768]: value = "task-1228709" [ 747.174250] env[61768]: _type = "Task" [ 747.174250] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.177926] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.178830] env[61768]: DEBUG nova.network.neutron [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Successfully created port: 45566649-d3e3-41d9-9059-90cd8cb8411d {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 747.186027] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228709, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.233020] env[61768]: DEBUG nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 747.334785] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228708, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061229} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.339441] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 747.340488] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3a499f-589b-4809-bcb8-897864659887 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.370473] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] 9fa32a4d-3ba4-4d36-963b-17a64453e804/9fa32a4d-3ba4-4d36-963b-17a64453e804.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 747.373358] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da17a3cb-d6c6-483e-9760-cd04a3eb96b8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.393734] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 747.393734] env[61768]: value = "task-1228710" [ 747.393734] env[61768]: _type = "Task" [ 747.393734] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.401475] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228710, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.473459] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 747.473756] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ee57400a-c77e-4114-8526-1c20dea6b91a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.482895] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 747.482895] env[61768]: value = "task-1228711" [ 747.482895] env[61768]: _type = "Task" [ 747.482895] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.493993] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fbf610-ee24-c9ca-dd8b-8dde7ebe9a74, 'name': SearchDatastore_Task, 'duration_secs': 0.023149} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.494963] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-633623ce-0d61-4a7b-9f44-c0cbf13640fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.501841] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228711, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.505239] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 747.505239] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528027a1-01d2-d1e5-338d-c0e6e97e0a64" [ 747.505239] env[61768]: _type = "Task" [ 747.505239] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.519011] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528027a1-01d2-d1e5-338d-c0e6e97e0a64, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.623627] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a0bbde-761c-4873-9cfc-de1fdecfddcd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.633383] env[61768]: DEBUG nova.network.neutron [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 747.634463] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c7082c-624e-40b6-86fe-2d8b9800ddd5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.669250] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a7e52f-38f2-4538-a53a-312cc7ca5177 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.683021] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6bd4ac-e455-4fd7-a753-67dbfef01004 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.695095] env[61768]: DEBUG nova.compute.provider_tree [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.700905] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228709, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.765966] env[61768]: DEBUG nova.network.neutron [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.904057] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.930134] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.930420] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.930644] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.930835] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.931026] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.933379] env[61768]: INFO nova.compute.manager [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Terminating instance [ 747.935358] env[61768]: DEBUG nova.compute.manager [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.935559] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 747.936628] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d84b73d1-2c68-4a39-a094-92438e5594f7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.945162] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 747.945498] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c0c091d-f96e-49ad-b9e0-b9341692826c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.952074] env[61768]: DEBUG oslo_vmware.api [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 747.952074] env[61768]: value = "task-1228712" [ 747.952074] env[61768]: _type = "Task" [ 747.952074] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.962787] env[61768]: DEBUG oslo_vmware.api [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228712, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.993444] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228711, 'name': PowerOffVM_Task, 'duration_secs': 0.220986} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.993720] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 747.993938] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 747.994720] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b54e40f-5dc8-4f02-be43-dbbdfa776ae0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.001653] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 748.001934] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-142a29db-a66a-4c03-b564-ead89ff9b7e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.017163] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528027a1-01d2-d1e5-338d-c0e6e97e0a64, 'name': SearchDatastore_Task, 'duration_secs': 0.016907} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.017407] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.017677] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 9b3632e5-4c2d-4968-9692-70f91883bfb3/9b3632e5-4c2d-4968-9692-70f91883bfb3.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 748.018534] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1f6770f-f3e7-402e-b74c-164be392583a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.024142] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 748.024142] env[61768]: value = "task-1228714" [ 748.024142] env[61768]: _type = "Task" [ 748.024142] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.033171] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228714, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.062054] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 748.062054] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 748.062172] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleting the datastore file [datastore1] efbde802-8cb0-4563-a776-3722a3889afe {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.062430] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51b918c0-e22d-4a14-aea8-f1eb6032b4e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.069235] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 748.069235] env[61768]: value = "task-1228715" [ 748.069235] env[61768]: _type = "Task" [ 748.069235] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.077474] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228715, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.185416] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228709, 'name': PowerOnVM_Task, 'duration_secs': 0.800239} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.185729] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 748.186022] env[61768]: INFO nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Took 10.15 seconds to spawn the instance on the hypervisor. [ 748.186315] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 748.187178] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f24059e-31c3-4b8a-a847-4aaf67434cb0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.198238] env[61768]: DEBUG nova.scheduler.client.report [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.244649] env[61768]: DEBUG nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 748.268395] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3bc9fc7-4aa1-47db-8eb4-791e4e1b7535 req-a0dd3c4f-01b8-4fe6-8477-7a09f0310a17 service nova] Releasing lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.268612] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquired lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.268774] env[61768]: DEBUG nova.network.neutron [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 748.273548] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 748.273809] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 748.273971] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 748.274204] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 748.274387] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 748.274569] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 748.274789] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 748.274956] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 748.275388] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 748.275593] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 748.276103] env[61768]: DEBUG nova.virt.hardware [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.277015] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8e3a2dd-6ea2-4963-8f1f-94d9d0ed5f7a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.286749] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09672f4a-aa64-414d-92c2-a3e2e838ae5a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.405974] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228710, 'name': ReconfigVM_Task, 'duration_secs': 0.551599} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.406334] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Reconfigured VM instance instance-00000035 to attach disk [datastore1] 9fa32a4d-3ba4-4d36-963b-17a64453e804/9fa32a4d-3ba4-4d36-963b-17a64453e804.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.407010] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3fa2799-009b-400f-b677-94dbd03e8514 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.422814] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 748.422814] env[61768]: value = "task-1228716" [ 748.422814] env[61768]: _type = "Task" [ 748.422814] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.435258] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228716, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.462653] env[61768]: DEBUG oslo_vmware.api [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228712, 'name': PowerOffVM_Task, 'duration_secs': 0.191713} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.462829] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 748.463012] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 748.463310] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ea284c2-c2a1-49a5-9d6b-736ca2faf42f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.530033] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 748.530033] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 748.530186] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Deleting the datastore file [datastore1] 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.530800] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4af80f2c-779b-49e5-ba43-04adff5f5a46 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.535344] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228714, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.472659} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.535897] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 9b3632e5-4c2d-4968-9692-70f91883bfb3/9b3632e5-4c2d-4968-9692-70f91883bfb3.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 748.536130] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 748.536355] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1a39398-475b-4fbf-9f9b-49a3560ff143 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.539518] env[61768]: DEBUG oslo_vmware.api [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for the task: (returnval){ [ 748.539518] env[61768]: value = "task-1228718" [ 748.539518] env[61768]: _type = "Task" [ 748.539518] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.543184] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 748.543184] env[61768]: value = "task-1228719" [ 748.543184] env[61768]: _type = "Task" [ 748.543184] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.549653] env[61768]: DEBUG oslo_vmware.api [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228718, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.554235] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228719, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.578625] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228715, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237984} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.578887] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.579214] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 748.579283] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 748.705558] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.489s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.706416] env[61768]: DEBUG nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 748.709774] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.077s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.710009] env[61768]: DEBUG nova.objects.instance [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lazy-loading 'resources' on Instance uuid 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.711140] env[61768]: INFO nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Took 29.23 seconds to build instance. [ 748.828267] env[61768]: DEBUG nova.network.neutron [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.888461] env[61768]: DEBUG nova.network.neutron [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Successfully updated port: 45566649-d3e3-41d9-9059-90cd8cb8411d {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 748.932206] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228716, 'name': Rename_Task, 'duration_secs': 0.20906} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.932687] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 748.932756] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d58e5a4d-63c6-4ff6-822a-f9dc608a37fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.940425] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 748.940425] env[61768]: value = "task-1228720" [ 748.940425] env[61768]: _type = "Task" [ 748.940425] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.949374] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228720, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.007080] env[61768]: DEBUG nova.network.neutron [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [{"id": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "address": "fa:16:3e:98:b7:f8", "network": {"id": "3e4a66df-17f2-479e-bb37-f8659f0180d6", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1225762599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ec5c061ca09040e8a2b3c3a6874daa40", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61e8c28f-c3", "ovs_interfaceid": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.052053] env[61768]: DEBUG oslo_vmware.api [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Task: {'id': task-1228718, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.451176} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.054745] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 749.054941] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 749.055203] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 749.055426] env[61768]: INFO nova.compute.manager [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Took 1.12 seconds to destroy the instance on the hypervisor. [ 749.055666] env[61768]: DEBUG oslo.service.loopingcall [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.056124] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228719, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072415} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.056331] env[61768]: DEBUG nova.compute.manager [-] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 749.056424] env[61768]: DEBUG nova.network.neutron [-] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 749.057918] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 749.058685] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b29c75b5-eaca-45be-b571-f5d60813712a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.081614] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Reconfiguring VM instance instance-00000036 to attach disk [datastore1] 9b3632e5-4c2d-4968-9692-70f91883bfb3/9b3632e5-4c2d-4968-9692-70f91883bfb3.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 749.081614] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-507edfcd-3319-4fbb-a585-effa7689d421 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.103683] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 749.103683] env[61768]: value = "task-1228721" [ 749.103683] env[61768]: _type = "Task" [ 749.103683] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.111191] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228721, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.117358] env[61768]: DEBUG nova.compute.manager [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Received event network-vif-plugged-45566649-d3e3-41d9-9059-90cd8cb8411d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.117582] env[61768]: DEBUG oslo_concurrency.lockutils [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] Acquiring lock "d70040f1-e425-4aeb-b82d-ce808a02a645-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.117820] env[61768]: DEBUG oslo_concurrency.lockutils [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.118010] env[61768]: DEBUG oslo_concurrency.lockutils [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.118215] env[61768]: DEBUG nova.compute.manager [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] No waiting events found dispatching network-vif-plugged-45566649-d3e3-41d9-9059-90cd8cb8411d {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 749.118395] env[61768]: WARNING nova.compute.manager [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Received unexpected event network-vif-plugged-45566649-d3e3-41d9-9059-90cd8cb8411d for instance with vm_state building and task_state spawning. [ 749.118562] env[61768]: DEBUG nova.compute.manager [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Received event network-changed-45566649-d3e3-41d9-9059-90cd8cb8411d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.118781] env[61768]: DEBUG nova.compute.manager [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Refreshing instance network info cache due to event network-changed-45566649-d3e3-41d9-9059-90cd8cb8411d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 749.118915] env[61768]: DEBUG oslo_concurrency.lockutils [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] Acquiring lock "refresh_cache-d70040f1-e425-4aeb-b82d-ce808a02a645" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.119049] env[61768]: DEBUG oslo_concurrency.lockutils [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] Acquired lock "refresh_cache-d70040f1-e425-4aeb-b82d-ce808a02a645" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.119235] env[61768]: DEBUG nova.network.neutron [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Refreshing network info cache for port 45566649-d3e3-41d9-9059-90cd8cb8411d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 749.216260] env[61768]: DEBUG nova.compute.utils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.221017] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.948s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.221017] env[61768]: DEBUG nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 749.221017] env[61768]: DEBUG nova.network.neutron [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 749.315928] env[61768]: DEBUG nova.policy [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '33db714fc1044225ac0286708a0bb7c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa6da53b22194f598ac23dae71b24c07', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 749.393804] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "refresh_cache-d70040f1-e425-4aeb-b82d-ce808a02a645" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.457862] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228720, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.510091] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Releasing lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.510091] env[61768]: DEBUG nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Instance network_info: |[{"id": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "address": "fa:16:3e:98:b7:f8", "network": {"id": "3e4a66df-17f2-479e-bb37-f8659f0180d6", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1225762599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ec5c061ca09040e8a2b3c3a6874daa40", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61e8c28f-c3", "ovs_interfaceid": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 749.511103] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:b7:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbf3349e-d05e-4d44-a011-c4b6e41af988', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61e8c28f-c389-4d6d-af8e-dabbc4c83363', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.519936] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Creating folder: Project (ec5c061ca09040e8a2b3c3a6874daa40). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 749.523061] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9be512a0-6416-486b-ac68-40040c683007 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.533936] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Created folder: Project (ec5c061ca09040e8a2b3c3a6874daa40) in parent group-v265360. [ 749.534164] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Creating folder: Instances. Parent ref: group-v265413. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 749.536541] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-564c5136-ea63-407e-9ca2-14ec3b2d8891 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.545485] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Created folder: Instances in parent group-v265413. [ 749.545719] env[61768]: DEBUG oslo.service.loopingcall [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.545913] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 749.546228] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c8a1502d-9546-464a-827d-4afff37508b9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.561550] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a83296b-fded-4a9b-b79d-f09f41c2a454 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.569254] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd9ca11-e930-4cba-99d0-4ff126c456cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.572117] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.572117] env[61768]: value = "task-1228724" [ 749.572117] env[61768]: _type = "Task" [ 749.572117] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.600641] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed6018e-0bc1-42a1-bd21-95182bd8fa54 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.608909] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228724, 'name': CreateVM_Task} progress is 15%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.616709] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca70c02-2cb5-4a59-8cc5-1faf0a0acb02 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.625493] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228721, 'name': ReconfigVM_Task, 'duration_secs': 0.296435} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.628430] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Reconfigured VM instance instance-00000036 to attach disk [datastore1] 9b3632e5-4c2d-4968-9692-70f91883bfb3/9b3632e5-4c2d-4968-9692-70f91883bfb3.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 749.629611] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-301c79e0-2ee2-4bf8-becb-2a7e3dbf1e93 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.639441] env[61768]: DEBUG nova.compute.provider_tree [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.642698] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 749.642940] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 749.643120] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 749.643432] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 749.643581] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 749.643740] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 749.643956] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 749.644239] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 749.644402] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 749.644579] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 749.644761] env[61768]: DEBUG nova.virt.hardware [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 749.646009] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb4b486-c4d2-40f9-8d5f-0b3bb29f2486 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.651402] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 749.651402] env[61768]: value = "task-1228725" [ 749.651402] env[61768]: _type = "Task" [ 749.651402] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.657697] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb8f60b-b707-4b9c-8fbf-8ea85e6cbddb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.668487] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228725, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.676451] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:02:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5379c85e-3ce3-433a-ad59-3b790a40b5fb', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 749.683685] env[61768]: DEBUG oslo.service.loopingcall [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 749.683999] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 749.684275] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9e35dee-90e9-4cc8-a135-3f217b1a52ba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.698584] env[61768]: DEBUG nova.network.neutron [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 749.703573] env[61768]: DEBUG nova.network.neutron [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Successfully created port: e2db4501-9ba5-4717-8ce1-2565bb20c603 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 749.708052] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 749.708052] env[61768]: value = "task-1228726" [ 749.708052] env[61768]: _type = "Task" [ 749.708052] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.714835] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228726, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.726722] env[61768]: DEBUG nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 749.729521] env[61768]: DEBUG nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.846162] env[61768]: DEBUG nova.network.neutron [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.953021] env[61768]: DEBUG oslo_vmware.api [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228720, 'name': PowerOnVM_Task, 'duration_secs': 0.962757} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.953373] env[61768]: DEBUG nova.network.neutron [-] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.954657] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 749.954859] env[61768]: INFO nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Took 9.49 seconds to spawn the instance on the hypervisor. [ 749.955209] env[61768]: DEBUG nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.955902] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42a2f8e-ae6a-4cf1-9fc6-ced52d9c9303 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.082353] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228724, 'name': CreateVM_Task, 'duration_secs': 0.330681} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.082576] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 750.083259] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.083443] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.083770] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 750.084056] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c8d9970-e08d-4e91-9df4-72cdd747a3d8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.088592] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 750.088592] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522e5f0b-2115-cded-af48-dd5eabdb8a8f" [ 750.088592] env[61768]: _type = "Task" [ 750.088592] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.095949] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522e5f0b-2115-cded-af48-dd5eabdb8a8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.150570] env[61768]: DEBUG nova.scheduler.client.report [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.162628] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228725, 'name': Rename_Task, 'duration_secs': 0.170664} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.162899] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 750.163149] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f1a9cfa0-22b3-478c-8ba0-6da3702024c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.169198] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 750.169198] env[61768]: value = "task-1228727" [ 750.169198] env[61768]: _type = "Task" [ 750.169198] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.176446] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228727, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.216951] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228726, 'name': CreateVM_Task, 'duration_secs': 0.391311} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.217085] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 750.217789] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.256763] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.349244] env[61768]: DEBUG oslo_concurrency.lockutils [req-cae37372-fbc8-467d-aecd-56a901d41e93 req-a314084c-f1ea-4678-886b-2c899d2a57bb service nova] Releasing lock "refresh_cache-d70040f1-e425-4aeb-b82d-ce808a02a645" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.349729] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquired lock "refresh_cache-d70040f1-e425-4aeb-b82d-ce808a02a645" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.349893] env[61768]: DEBUG nova.network.neutron [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 750.460629] env[61768]: INFO nova.compute.manager [-] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Took 1.40 seconds to deallocate network for instance. [ 750.480947] env[61768]: INFO nova.compute.manager [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Took 27.27 seconds to build instance. [ 750.601867] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522e5f0b-2115-cded-af48-dd5eabdb8a8f, 'name': SearchDatastore_Task, 'duration_secs': 0.011209} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.602206] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.602582] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.603224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 750.603224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.603224] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.603402] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.604337] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 750.604337] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f47b20f6-020f-48d2-a70f-e4226fcb48a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.605905] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0da6d99f-653f-4e5e-a169-11fd446027ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.611143] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 750.611143] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526a8de5-3018-2c0a-d305-57e00ca377b0" [ 750.611143] env[61768]: _type = "Task" [ 750.611143] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.616075] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 750.616075] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 750.617216] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-267ff5b8-0ab9-42a3-a091-1d8428a9300c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.622681] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526a8de5-3018-2c0a-d305-57e00ca377b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.626159] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 750.626159] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525f29e1-bb13-390d-0c0a-1502b98b1ced" [ 750.626159] env[61768]: _type = "Task" [ 750.626159] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.634059] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525f29e1-bb13-390d-0c0a-1502b98b1ced, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.655947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.946s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.659233] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.669s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.659844] env[61768]: DEBUG nova.objects.instance [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lazy-loading 'resources' on Instance uuid 3401be45-c3ab-4780-81a7-a93b05742414 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 750.681450] env[61768]: DEBUG oslo_vmware.api [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228727, 'name': PowerOnVM_Task, 'duration_secs': 0.495339} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.681450] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 750.681594] env[61768]: INFO nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Took 7.30 seconds to spawn the instance on the hypervisor. [ 750.682661] env[61768]: DEBUG nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 750.684771] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ceae42-c88a-4121-8c60-19c924c45b21 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.698070] env[61768]: INFO nova.scheduler.client.report [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Deleted allocations for instance 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed [ 750.738396] env[61768]: DEBUG nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 750.771087] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 750.771087] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 750.771087] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 750.771275] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 750.771459] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 750.771627] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 750.771843] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 750.772013] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 750.772193] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 750.772360] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 750.772536] env[61768]: DEBUG nova.virt.hardware [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.773709] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2284b2a9-3eaf-4e9f-884f-af7d7703264b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.782445] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3bf2fed-d31a-4afe-b82e-5712c2a7f6dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.896942] env[61768]: DEBUG nova.network.neutron [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.978943] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.983528] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3d687ad6-b798-47bd-a6e6-e97794853d43 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.688s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.078549] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.078837] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.079745] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.079911] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.080110] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.082632] env[61768]: INFO nova.compute.manager [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Terminating instance [ 751.084534] env[61768]: DEBUG nova.compute.manager [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.084738] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.085566] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab91472b-bd7c-47d1-8e4b-b483c3f94ae0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.089922] env[61768]: DEBUG nova.network.neutron [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Updating instance_info_cache with network_info: [{"id": "45566649-d3e3-41d9-9059-90cd8cb8411d", "address": "fa:16:3e:33:63:95", "network": {"id": "cec072ef-1454-43fa-bd6e-0a1aa530e1d9", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1537542821-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "272b4bfd904e4bfebb692b0492f955c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45566649-d3", "ovs_interfaceid": "45566649-d3e3-41d9-9059-90cd8cb8411d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.096478] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 751.096722] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6dbd67f-a224-438e-b275-f4468d1ea947 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.103853] env[61768]: DEBUG oslo_vmware.api [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 751.103853] env[61768]: value = "task-1228728" [ 751.103853] env[61768]: _type = "Task" [ 751.103853] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.113276] env[61768]: DEBUG oslo_vmware.api [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.122279] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526a8de5-3018-2c0a-d305-57e00ca377b0, 'name': SearchDatastore_Task, 'duration_secs': 0.014499} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.122587] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.122823] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 751.123054] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.134346] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525f29e1-bb13-390d-0c0a-1502b98b1ced, 'name': SearchDatastore_Task, 'duration_secs': 0.014811} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.135134] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39daa82a-c6d2-46da-bdd2-6f09fe472047 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.140620] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 751.140620] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525b8f51-e360-5173-d2e7-ae04e7f85489" [ 751.140620] env[61768]: _type = "Task" [ 751.140620] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.148253] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525b8f51-e360-5173-d2e7-ae04e7f85489, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.160055] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "9fa32a4d-3ba4-4d36-963b-17a64453e804" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.160206] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.160445] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "9fa32a4d-3ba4-4d36-963b-17a64453e804-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.160657] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.160855] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.165569] env[61768]: INFO nova.compute.manager [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Terminating instance [ 751.167696] env[61768]: DEBUG nova.compute.manager [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 751.167921] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 751.168792] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f39bb852-a67e-410b-805e-cd80e913be28 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.173781] env[61768]: DEBUG nova.compute.manager [req-b4ce15e5-f323-49ea-9cd2-311f4f3ece9a req-eb321c13-2e83-4c53-801c-03e66cfc44f4 service nova] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Received event network-vif-deleted-79d4c508-cec4-40c8-8974-b1916ac2a4d2 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.179836] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 751.180255] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08192367-6217-44d6-9b20-dd79fe94474e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.187872] env[61768]: DEBUG oslo_vmware.api [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 751.187872] env[61768]: value = "task-1228729" [ 751.187872] env[61768]: _type = "Task" [ 751.187872] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.208349] env[61768]: DEBUG oslo_vmware.api [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228729, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.210659] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6482fbf8-c07b-42ab-b81b-7aa70f00add2 tempest-SecurityGroupsTestJSON-1379887877 tempest-SecurityGroupsTestJSON-1379887877-project-member] Lock "7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.512s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.222030] env[61768]: INFO nova.compute.manager [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Took 27.92 seconds to build instance. [ 751.455170] env[61768]: DEBUG nova.network.neutron [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Successfully updated port: e2db4501-9ba5-4717-8ce1-2565bb20c603 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 751.473969] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee40d725-0c2d-4938-9a88-042180daa4c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.482086] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f2e940-6d32-4603-ab38-3a6d3b01dd5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.485449] env[61768]: DEBUG nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 751.520725] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea0f89a-679c-4438-9dfe-a65495ab76fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.530324] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26617227-f302-411d-8a7e-7f3b75777552 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.546025] env[61768]: DEBUG nova.compute.provider_tree [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.591947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Releasing lock "refresh_cache-d70040f1-e425-4aeb-b82d-ce808a02a645" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.592323] env[61768]: DEBUG nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Instance network_info: |[{"id": "45566649-d3e3-41d9-9059-90cd8cb8411d", "address": "fa:16:3e:33:63:95", "network": {"id": "cec072ef-1454-43fa-bd6e-0a1aa530e1d9", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1537542821-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "272b4bfd904e4bfebb692b0492f955c8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bfae3ef8-cae7-455d-8632-ba93e1671625", "external-id": "cl2-zone-841", "segmentation_id": 841, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45566649-d3", "ovs_interfaceid": "45566649-d3e3-41d9-9059-90cd8cb8411d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 751.592772] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:33:63:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bfae3ef8-cae7-455d-8632-ba93e1671625', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45566649-d3e3-41d9-9059-90cd8cb8411d', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.600265] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Creating folder: Project (272b4bfd904e4bfebb692b0492f955c8). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 751.600544] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8bece1b-0ac5-4f4f-8528-473fbd8bd9c7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.611806] env[61768]: DEBUG oslo_vmware.api [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228728, 'name': PowerOffVM_Task, 'duration_secs': 0.205195} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.612986] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 751.613182] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 751.613440] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Created folder: Project (272b4bfd904e4bfebb692b0492f955c8) in parent group-v265360. [ 751.613611] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Creating folder: Instances. Parent ref: group-v265417. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 751.613825] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ae894a12-bdff-48c4-b58e-000fb9c62bb7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.615265] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ae3fb27-2cc9-4e39-a4ad-8fedf7628bc3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.624522] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Created folder: Instances in parent group-v265417. [ 751.624757] env[61768]: DEBUG oslo.service.loopingcall [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.624941] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 751.625184] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-eeb5116c-addf-49c1-8e82-b1b15f473a2d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.645104] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.645104] env[61768]: value = "task-1228733" [ 751.645104] env[61768]: _type = "Task" [ 751.645104] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.651527] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525b8f51-e360-5173-d2e7-ae04e7f85489, 'name': SearchDatastore_Task, 'duration_secs': 0.011044} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.652144] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.652411] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/23a4450d-162a-4b2b-a009-7023851315a1.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 751.652704] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.652896] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 751.653116] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2842d4b7-af7c-4537-92bc-1e01861047c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.657944] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c9f5e2e7-fa28-4eb0-9646-f2348b6788b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.659578] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228733, 'name': CreateVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.663484] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 751.663484] env[61768]: value = "task-1228734" [ 751.663484] env[61768]: _type = "Task" [ 751.663484] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.667952] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 751.668191] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 751.671359] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2356579-60e7-43dc-a9d1-fb2697aa0b1b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.673969] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228734, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.676405] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 751.676731] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 751.676935] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleting the datastore file [datastore1] 668d4fe4-85f8-4282-8d65-0549d9e3bda8 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.678276] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f6b7a6e-3eaf-494b-8110-679891035ea6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.680223] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 751.680223] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52706c6f-d92b-587e-d3c7-226468d32fcb" [ 751.680223] env[61768]: _type = "Task" [ 751.680223] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.685400] env[61768]: DEBUG oslo_vmware.api [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 751.685400] env[61768]: value = "task-1228735" [ 751.685400] env[61768]: _type = "Task" [ 751.685400] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.692614] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52706c6f-d92b-587e-d3c7-226468d32fcb, 'name': SearchDatastore_Task, 'duration_secs': 0.010797} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.698026] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19d9a97e-f475-41e8-b1e7-3bc07b2dffee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.703975] env[61768]: DEBUG oslo_vmware.api [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228735, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.708013] env[61768]: DEBUG oslo_vmware.api [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228729, 'name': PowerOffVM_Task, 'duration_secs': 0.201673} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.709300] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 751.709545] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 751.712664] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 751.712664] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52345256-138c-41a8-9a1e-ac8077d074db" [ 751.712664] env[61768]: _type = "Task" [ 751.712664] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.712664] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1cd1a48a-2512-4ab7-8e8e-d750e907b59e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.720482] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52345256-138c-41a8-9a1e-ac8077d074db, 'name': SearchDatastore_Task, 'duration_secs': 0.010153} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.720738] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.722677] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 751.722677] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86ad8a38-cf12-409d-9647-d2ac87288803 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.723627] env[61768]: DEBUG oslo_concurrency.lockutils [None req-43c81d59-0fbe-41c8-add5-a1b75ce0652c tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.836s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.727542] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 751.727542] env[61768]: value = "task-1228737" [ 751.727542] env[61768]: _type = "Task" [ 751.727542] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.735399] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228737, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.773737] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 751.773737] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 751.773737] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleting the datastore file [datastore1] 9fa32a4d-3ba4-4d36-963b-17a64453e804 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 751.773737] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55a97afc-161f-4200-8e28-2ac39c661b4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.781385] env[61768]: DEBUG oslo_vmware.api [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 751.781385] env[61768]: value = "task-1228738" [ 751.781385] env[61768]: _type = "Task" [ 751.781385] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.789651] env[61768]: DEBUG oslo_vmware.api [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228738, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.959861] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "refresh_cache-9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.959861] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "refresh_cache-9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.960165] env[61768]: DEBUG nova.network.neutron [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.008804] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.048836] env[61768]: DEBUG nova.scheduler.client.report [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.157378] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228733, 'name': CreateVM_Task, 'duration_secs': 0.350147} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.157589] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 752.158379] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.158573] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.158904] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.159198] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f20c6d68-a03f-4b32-bb21-5da94e866a6c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.164604] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 752.164604] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527bbd13-9a8f-07e3-af0a-f161b2d2d429" [ 752.164604] env[61768]: _type = "Task" [ 752.164604] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.176370] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228734, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465617} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.179901] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/23a4450d-162a-4b2b-a009-7023851315a1.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 752.180243] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 752.181244] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527bbd13-9a8f-07e3-af0a-f161b2d2d429, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.181244] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5a307a8c-980d-4132-ae05-110183d5d836 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.187712] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 752.187712] env[61768]: value = "task-1228739" [ 752.187712] env[61768]: _type = "Task" [ 752.187712] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.200678] env[61768]: DEBUG oslo_vmware.api [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228735, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190392} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.203721] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.203809] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 752.203986] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.204237] env[61768]: INFO nova.compute.manager [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 752.204513] env[61768]: DEBUG oslo.service.loopingcall [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.205178] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228739, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.205411] env[61768]: DEBUG nova.compute.manager [-] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.206032] env[61768]: DEBUG nova.network.neutron [-] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.226466] env[61768]: DEBUG nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.238941] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228737, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.291444] env[61768]: DEBUG oslo_vmware.api [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228738, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.3169} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.291832] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.291922] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 752.292220] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 752.292449] env[61768]: INFO nova.compute.manager [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Took 1.12 seconds to destroy the instance on the hypervisor. [ 752.292797] env[61768]: DEBUG oslo.service.loopingcall [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.292943] env[61768]: DEBUG nova.compute.manager [-] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.293110] env[61768]: DEBUG nova.network.neutron [-] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 752.295468] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "9b3632e5-4c2d-4968-9692-70f91883bfb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.295746] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.296177] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "9b3632e5-4c2d-4968-9692-70f91883bfb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.296177] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.296360] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.299473] env[61768]: INFO nova.compute.manager [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Terminating instance [ 752.301533] env[61768]: DEBUG nova.compute.manager [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 752.301748] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 752.302613] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-497653a7-c401-486b-aa9e-f48b29bd9fdd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.312557] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 752.312835] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-21c85ff2-2b7f-409c-ab68-621621e7c9e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.320144] env[61768]: DEBUG oslo_vmware.api [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 752.320144] env[61768]: value = "task-1228740" [ 752.320144] env[61768]: _type = "Task" [ 752.320144] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.329965] env[61768]: DEBUG oslo_vmware.api [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.518849] env[61768]: DEBUG nova.network.neutron [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.554083] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.895s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.559152] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.269s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.560641] env[61768]: INFO nova.compute.claims [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.582590] env[61768]: INFO nova.scheduler.client.report [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Deleted allocations for instance 3401be45-c3ab-4780-81a7-a93b05742414 [ 752.678253] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527bbd13-9a8f-07e3-af0a-f161b2d2d429, 'name': SearchDatastore_Task, 'duration_secs': 0.053213} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.678617] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.678863] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.679134] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.679282] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.683058] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.683058] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a314c55b-e222-4185-9d88-234e4d07e73a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.687890] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.688084] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 752.690218] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9417ce89-135c-4d7a-9465-d19c672a7d8f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.704759] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 752.704759] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525000b2-0856-0591-c7b7-3f63b0b0c994" [ 752.704759] env[61768]: _type = "Task" [ 752.704759] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.709734] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228739, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099736} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.713673] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.714475] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34ca7b2-c5f0-4db2-868c-6793f8fb201f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.722800] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525000b2-0856-0591-c7b7-3f63b0b0c994, 'name': SearchDatastore_Task, 'duration_secs': 0.010095} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.744676] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/23a4450d-162a-4b2b-a009-7023851315a1.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.745821] env[61768]: DEBUG nova.network.neutron [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Updating instance_info_cache with network_info: [{"id": "e2db4501-9ba5-4717-8ce1-2565bb20c603", "address": "fa:16:3e:07:c4:a4", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2db4501-9b", "ovs_interfaceid": "e2db4501-9ba5-4717-8ce1-2565bb20c603", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.751029] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03ecc8cb-01d9-4f10-a20a-40048d6bc504 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.754401] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3801e78e-64c3-4879-a886-c84e6e175e68 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.777698] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 752.777698] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523b04a6-48b7-58db-e17a-348359159f23" [ 752.777698] env[61768]: _type = "Task" [ 752.777698] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.783192] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228737, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.698236} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.786934] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 752.787045] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 752.787602] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 752.787602] env[61768]: value = "task-1228741" [ 752.787602] env[61768]: _type = "Task" [ 752.787602] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.788320] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.790408] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac37e5e1-bb2c-49d1-846b-f58aee6b66a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.802023] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523b04a6-48b7-58db-e17a-348359159f23, 'name': SearchDatastore_Task, 'duration_secs': 0.009993} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.802023] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.802023] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] d70040f1-e425-4aeb-b82d-ce808a02a645/d70040f1-e425-4aeb-b82d-ce808a02a645.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 752.802461] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef0fbd8e-b28f-4eac-98e6-6e3a807e0e67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.810029] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228741, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.810451] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 752.810451] env[61768]: value = "task-1228742" [ 752.810451] env[61768]: _type = "Task" [ 752.810451] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.816294] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 752.816294] env[61768]: value = "task-1228743" [ 752.816294] env[61768]: _type = "Task" [ 752.816294] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.826321] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228742, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.834808] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228743, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.837886] env[61768]: DEBUG oslo_vmware.api [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228740, 'name': PowerOffVM_Task, 'duration_secs': 0.198919} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.838197] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 752.838409] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 752.838671] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de8cdd5f-6174-4527-b8ab-dbff4341c3ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.914103] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 752.914103] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 752.914103] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Deleting the datastore file [datastore1] 9b3632e5-4c2d-4968-9692-70f91883bfb3 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 752.914103] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2f446567-a348-48d2-9782-c4396533b8af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.920252] env[61768]: DEBUG oslo_vmware.api [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for the task: (returnval){ [ 752.920252] env[61768]: value = "task-1228745" [ 752.920252] env[61768]: _type = "Task" [ 752.920252] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.928347] env[61768]: DEBUG oslo_vmware.api [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228745, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.090447] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0cff313e-4cd8-4963-8cac-780d0f7ff71f tempest-ServerMetadataTestJSON-498033897 tempest-ServerMetadataTestJSON-498033897-project-member] Lock "3401be45-c3ab-4780-81a7-a93b05742414" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.180s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.247760] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Received event network-vif-plugged-e2db4501-9ba5-4717-8ce1-2565bb20c603 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.248207] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] Acquiring lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.248364] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.248598] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.248812] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] No waiting events found dispatching network-vif-plugged-e2db4501-9ba5-4717-8ce1-2565bb20c603 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 753.249023] env[61768]: WARNING nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Received unexpected event network-vif-plugged-e2db4501-9ba5-4717-8ce1-2565bb20c603 for instance with vm_state building and task_state spawning. [ 753.249227] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Received event network-changed-e2db4501-9ba5-4717-8ce1-2565bb20c603 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.249407] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Refreshing instance network info cache due to event network-changed-e2db4501-9ba5-4717-8ce1-2565bb20c603. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 753.249625] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] Acquiring lock "refresh_cache-9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.273586] env[61768]: DEBUG nova.network.neutron [-] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.275705] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "refresh_cache-9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.276140] env[61768]: DEBUG nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance network_info: |[{"id": "e2db4501-9ba5-4717-8ce1-2565bb20c603", "address": "fa:16:3e:07:c4:a4", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2db4501-9b", "ovs_interfaceid": "e2db4501-9ba5-4717-8ce1-2565bb20c603", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 753.276877] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] Acquired lock "refresh_cache-9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.277271] env[61768]: DEBUG nova.network.neutron [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Refreshing network info cache for port e2db4501-9ba5-4717-8ce1-2565bb20c603 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 753.278944] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:c4:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2db4501-9ba5-4717-8ce1-2565bb20c603', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.291597] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating folder: Project (fa6da53b22194f598ac23dae71b24c07). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 753.297770] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-633612e4-28b4-400d-9d8e-8b265c3b9bdb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.305071] env[61768]: DEBUG nova.network.neutron [-] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.312766] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228741, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.318142] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created folder: Project (fa6da53b22194f598ac23dae71b24c07) in parent group-v265360. [ 753.318142] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating folder: Instances. Parent ref: group-v265420. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 753.321177] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8620386b-5b31-44e5-86df-2e632ce8b449 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.326228] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228742, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072299} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.327316] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.328190] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0fda53-e360-4e59-9c83-ec64d0208762 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.333743] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228743, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.506203} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.338083] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] d70040f1-e425-4aeb-b82d-ce808a02a645/d70040f1-e425-4aeb-b82d-ce808a02a645.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 753.338083] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 753.338083] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created folder: Instances in parent group-v265420. [ 753.338083] env[61768]: DEBUG oslo.service.loopingcall [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.338083] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1d08f62b-9d98-488d-bbc5-b383c05a9e14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.338395] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 753.348114] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b6dde4e5-7f70-45ee-9cbc-84600652cccf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.370940] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.374251] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-929c66c3-fc4b-46c0-bdc1-42cf70109cdd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.392389] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 753.392389] env[61768]: value = "task-1228748" [ 753.392389] env[61768]: _type = "Task" [ 753.392389] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.397231] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.397231] env[61768]: value = "task-1228749" [ 753.397231] env[61768]: _type = "Task" [ 753.397231] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.399187] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 753.399187] env[61768]: value = "task-1228750" [ 753.399187] env[61768]: _type = "Task" [ 753.399187] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.411463] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228748, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.417067] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228749, 'name': CreateVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.420445] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228750, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.434459] env[61768]: DEBUG oslo_vmware.api [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Task: {'id': task-1228745, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.319714} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.434748] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 753.434949] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 753.435154] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 753.435336] env[61768]: INFO nova.compute.manager [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Took 1.13 seconds to destroy the instance on the hypervisor. [ 753.435625] env[61768]: DEBUG oslo.service.loopingcall [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.435826] env[61768]: DEBUG nova.compute.manager [-] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.435919] env[61768]: DEBUG nova.network.neutron [-] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.758564] env[61768]: DEBUG nova.network.neutron [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Updated VIF entry in instance network info cache for port e2db4501-9ba5-4717-8ce1-2565bb20c603. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 753.758973] env[61768]: DEBUG nova.network.neutron [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Updating instance_info_cache with network_info: [{"id": "e2db4501-9ba5-4717-8ce1-2565bb20c603", "address": "fa:16:3e:07:c4:a4", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape2db4501-9b", "ovs_interfaceid": "e2db4501-9ba5-4717-8ce1-2565bb20c603", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.780279] env[61768]: INFO nova.compute.manager [-] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Took 1.48 seconds to deallocate network for instance. [ 753.809910] env[61768]: INFO nova.compute.manager [-] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Took 1.60 seconds to deallocate network for instance. [ 753.810135] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228741, 'name': ReconfigVM_Task, 'duration_secs': 0.542252} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.811983] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/23a4450d-162a-4b2b-a009-7023851315a1.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 753.815820] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ed4cefc-d90c-4169-9f42-cbd79704f429 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.825130] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 753.825130] env[61768]: value = "task-1228751" [ 753.825130] env[61768]: _type = "Task" [ 753.825130] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.834413] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228751, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.906419] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228748, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070484} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.911260] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 753.911948] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f69421-a1cc-429b-a3f8-1c3ed3a36d72 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.918136] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75bd1a04-4520-4a5f-998d-53773154bcc4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.920457] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228749, 'name': CreateVM_Task, 'duration_secs': 0.388536} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.921781] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 753.923204] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.923426] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.923811] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 753.937890] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3015789b-f075-4d04-a73a-19aae60f0cdc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.949158] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] d70040f1-e425-4aeb-b82d-ce808a02a645/d70040f1-e425-4aeb-b82d-ce808a02a645.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 753.949547] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228750, 'name': ReconfigVM_Task, 'duration_secs': 0.408189} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.953457] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed0d992-3946-4f45-b807-d2e2e4c41fec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.959105] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-726ce862-ab03-46d9-ae08-a0c4d70e5b60 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.979700] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Reconfigured VM instance instance-00000032 to attach disk [datastore2] efbde802-8cb0-4563-a776-3722a3889afe/efbde802-8cb0-4563-a776-3722a3889afe.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 753.981431] env[61768]: DEBUG nova.network.neutron [-] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.983586] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc404054-9044-42cf-be36-e278398a0b51 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.986720] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 753.986720] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c7744c-d644-a1f1-f2c9-1b5d8c14216d" [ 753.986720] env[61768]: _type = "Task" [ 753.986720] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.023139] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 754.023139] env[61768]: value = "task-1228752" [ 754.023139] env[61768]: _type = "Task" [ 754.023139] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.023139] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 754.023139] env[61768]: value = "task-1228753" [ 754.023139] env[61768]: _type = "Task" [ 754.023139] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.023139] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-432bf79a-caf2-471c-afe1-7d2b89125e30 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.035794] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c7744c-d644-a1f1-f2c9-1b5d8c14216d, 'name': SearchDatastore_Task, 'duration_secs': 0.013455} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.036784] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.037141] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.037507] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.037761] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.038085] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.040822] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8480103-0546-47d0-b813-1bb02b52d973 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.050118] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba7bff78-afd4-4297-a347-4a1c6106157b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.053659] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228753, 'name': Rename_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.053966] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.066972] env[61768]: DEBUG nova.compute.provider_tree [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.067211] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.067419] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 754.068326] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac763f96-ec7a-447d-aa9c-9d0a0d318efa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.077023] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 754.077023] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5232b7ce-7735-119f-2d7e-9ba409f6b1c7" [ 754.077023] env[61768]: _type = "Task" [ 754.077023] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.082604] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5232b7ce-7735-119f-2d7e-9ba409f6b1c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.262328] env[61768]: DEBUG oslo_concurrency.lockutils [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] Releasing lock "refresh_cache-9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.262610] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Received event network-vif-deleted-16c13000-f458-4c8b-be59-ea8c7c28d9fc {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.262793] env[61768]: INFO nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Neutron deleted interface 16c13000-f458-4c8b-be59-ea8c7c28d9fc; detaching it from the instance and deleting it from the info cache [ 754.262968] env[61768]: DEBUG nova.network.neutron [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.288445] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.321136] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.336173] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228751, 'name': Rename_Task, 'duration_secs': 0.145823} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.336173] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 754.336173] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8c9e5b7-7d84-4264-87e0-fe6f20d9cdc7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.341883] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 754.341883] env[61768]: value = "task-1228754" [ 754.341883] env[61768]: _type = "Task" [ 754.341883] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.351875] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228754, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.487380] env[61768]: INFO nova.compute.manager [-] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Took 1.05 seconds to deallocate network for instance. [ 754.539473] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228753, 'name': Rename_Task, 'duration_secs': 0.141791} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.542708] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 754.543278] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.543507] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4916b5e7-5fe7-480a-85ac-87bfc7c3cc9d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.549764] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 754.549764] env[61768]: value = "task-1228755" [ 754.549764] env[61768]: _type = "Task" [ 754.549764] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.557340] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228755, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.569772] env[61768]: DEBUG nova.scheduler.client.report [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.586452] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5232b7ce-7735-119f-2d7e-9ba409f6b1c7, 'name': SearchDatastore_Task, 'duration_secs': 0.009305} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.587868] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa130f0d-b211-48ca-aab9-e19464d2aa93 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.594440] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 754.594440] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527412b6-787c-49f3-6049-a1e62973aacc" [ 754.594440] env[61768]: _type = "Task" [ 754.594440] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.604369] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527412b6-787c-49f3-6049-a1e62973aacc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.766726] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-163b86f9-5ec0-4066-9068-5c5a5d1191f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.777744] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d968eb-b9c5-40d7-9f1e-11d4212ad03b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.804380] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Detach interface failed, port_id=16c13000-f458-4c8b-be59-ea8c7c28d9fc, reason: Instance 9fa32a4d-3ba4-4d36-963b-17a64453e804 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 754.804716] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Received event network-vif-deleted-3c8dc29a-e757-4d7d-93a4-9274629ec62c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 754.804899] env[61768]: INFO nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Neutron deleted interface 3c8dc29a-e757-4d7d-93a4-9274629ec62c; detaching it from the instance and deleting it from the info cache [ 754.805319] env[61768]: DEBUG nova.network.neutron [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.853974] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228754, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.993191] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.039700] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228752, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.059610] env[61768]: DEBUG oslo_vmware.api [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228755, 'name': PowerOnVM_Task, 'duration_secs': 0.469357} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.059892] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 755.060264] env[61768]: DEBUG nova.compute.manager [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.061066] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0801c75-a0d4-46dc-abcf-56d248955dae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.074418] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.074869] env[61768]: DEBUG nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 755.077662] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 21.518s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 755.104059] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527412b6-787c-49f3-6049-a1e62973aacc, 'name': SearchDatastore_Task, 'duration_secs': 0.011598} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.104366] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.105098] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 755.105458] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9b44cad2-1443-4d66-9ecb-22c33456bc5b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.113449] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 755.113449] env[61768]: value = "task-1228756" [ 755.113449] env[61768]: _type = "Task" [ 755.113449] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.122269] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228756, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.297046] env[61768]: DEBUG nova.compute.manager [req-937374ee-554a-4001-8225-ecf72edaa2d3 req-2f5cffa1-ce2e-4bc1-8487-e84373f46807 service nova] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Received event network-vif-deleted-be7be29f-b1a1-4851-b1a4-b368519c25ef {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.315049] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ad272698-44f1-4e8b-8903-12b5fae6df89 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.324031] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9dbb3f-0bd8-4cc2-b736-b035cfb0ee5b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.352147] env[61768]: DEBUG nova.compute.manager [req-0ba3b65e-ac15-4873-b3f5-077b46d024e4 req-25255be2-d249-4cf7-bbcf-36cfedd09bd4 service nova] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Detach interface failed, port_id=3c8dc29a-e757-4d7d-93a4-9274629ec62c, reason: Instance 668d4fe4-85f8-4282-8d65-0549d9e3bda8 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 755.362703] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228754, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.539303] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228752, 'name': ReconfigVM_Task, 'duration_secs': 1.229082} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.539303] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Reconfigured VM instance instance-00000038 to attach disk [datastore2] d70040f1-e425-4aeb-b82d-ce808a02a645/d70040f1-e425-4aeb-b82d-ce808a02a645.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 755.539303] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af249c49-fc8d-4c28-8150-ded5b42e4000 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.551570] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 755.551570] env[61768]: value = "task-1228757" [ 755.551570] env[61768]: _type = "Task" [ 755.551570] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.564092] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228757, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.579123] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.591056] env[61768]: DEBUG nova.compute.utils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.596022] env[61768]: DEBUG nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.596022] env[61768]: DEBUG nova.network.neutron [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 755.624193] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228756, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.643127] env[61768]: DEBUG nova.policy [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '33db714fc1044225ac0286708a0bb7c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa6da53b22194f598ac23dae71b24c07', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.862790] env[61768]: DEBUG oslo_vmware.api [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228754, 'name': PowerOnVM_Task, 'duration_secs': 1.043856} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.863118] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 755.863639] env[61768]: INFO nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Took 9.99 seconds to spawn the instance on the hypervisor. [ 755.863639] env[61768]: DEBUG nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.864266] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-905e580c-5aef-4a53-b9bd-e3149e7ccf57 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.925942] env[61768]: DEBUG nova.network.neutron [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Successfully created port: a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 756.063021] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228757, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.098356] env[61768]: DEBUG nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 756.124184] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228756, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588069} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.124514] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 756.124752] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 756.125050] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-797cf093-9e5a-4990-ab8c-41be874764ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.131407] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 756.131407] env[61768]: value = "task-1228758" [ 756.131407] env[61768]: _type = "Task" [ 756.131407] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.139920] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228758, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.277047] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 12a543c2-1081-49d7-800b-07f0a2516904 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 756.277286] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 756.277442] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 0df161e9-4a66-432e-9366-60f231c53e3d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 756.278284] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 91d56a8c-0664-4b3c-bf28-e2c668ac65fa is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 756.278468] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance efbde802-8cb0-4563-a776-3722a3889afe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 756.278611] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 756.278761] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 668d4fe4-85f8-4282-8d65-0549d9e3bda8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 756.278885] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 9fa32a4d-3ba4-4d36-963b-17a64453e804 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 756.279011] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 9b3632e5-4c2d-4968-9692-70f91883bfb3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 756.279137] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 23a4450d-162a-4b2b-a009-7023851315a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 756.279468] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance d70040f1-e425-4aeb-b82d-ce808a02a645 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 756.279468] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 756.279584] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 756.380397] env[61768]: INFO nova.compute.manager [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Took 28.07 seconds to build instance. [ 756.566063] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228757, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.644314] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228758, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.094306} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.644612] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.645815] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196875b3-6add-456b-b1b9-63a38514d8ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.668017] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.668332] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9ca59e8-de50-4687-ae65-6c96519868aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.684329] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "efbde802-8cb0-4563-a776-3722a3889afe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.684585] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "efbde802-8cb0-4563-a776-3722a3889afe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.684871] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "efbde802-8cb0-4563-a776-3722a3889afe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.684981] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "efbde802-8cb0-4563-a776-3722a3889afe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.685171] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "efbde802-8cb0-4563-a776-3722a3889afe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.687491] env[61768]: INFO nova.compute.manager [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Terminating instance [ 756.689281] env[61768]: DEBUG nova.compute.manager [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 756.689480] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 756.690288] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f192def-b1f4-4b45-a036-c8a439672ef2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.695042] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 756.695042] env[61768]: value = "task-1228759" [ 756.695042] env[61768]: _type = "Task" [ 756.695042] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.700611] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 756.701214] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-090b17e2-a22a-4278-b3f3-aa70644da754 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.705541] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228759, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.709542] env[61768]: DEBUG oslo_vmware.api [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 756.709542] env[61768]: value = "task-1228760" [ 756.709542] env[61768]: _type = "Task" [ 756.709542] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.719713] env[61768]: DEBUG oslo_vmware.api [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228760, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.782632] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 76c282d1-ddfe-46dc-aa7b-225708443379 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 756.882391] env[61768]: DEBUG oslo_concurrency.lockutils [None req-94525ce7-0e40-4f86-8ccd-d5570a8b5df4 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "23a4450d-162a-4b2b-a009-7023851315a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.395s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.974538] env[61768]: INFO nova.compute.manager [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Rescuing [ 756.974538] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.974538] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquired lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.974538] env[61768]: DEBUG nova.network.neutron [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 757.064047] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228757, 'name': Rename_Task, 'duration_secs': 1.349102} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.064663] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 757.064745] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b8a85d1-b18e-4872-95bb-26a60f3c3e63 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.076019] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 757.076019] env[61768]: value = "task-1228761" [ 757.076019] env[61768]: _type = "Task" [ 757.076019] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.084146] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228761, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.108603] env[61768]: DEBUG nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 757.140552] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 757.140861] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 757.141044] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 757.141435] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 757.141669] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 757.141838] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 757.142067] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 757.142239] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 757.142438] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 757.142679] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 757.142830] env[61768]: DEBUG nova.virt.hardware [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.143723] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9035c53e-e813-405d-b186-79fa5cb24f75 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.153072] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09050e07-ad3c-4716-b804-d029275d1f22 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.205134] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228759, 'name': ReconfigVM_Task, 'duration_secs': 0.389595} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.205509] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.206242] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65eca66b-7c17-41cb-b402-1a062c8ead30 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.214550] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 757.214550] env[61768]: value = "task-1228762" [ 757.214550] env[61768]: _type = "Task" [ 757.214550] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.222034] env[61768]: DEBUG oslo_vmware.api [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228760, 'name': PowerOffVM_Task, 'duration_secs': 0.257958} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.222670] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 757.222848] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 757.223115] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-51a93c7b-c72f-4f1f-9e14-7d01976c23ee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.228584] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228762, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.280862] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "4f429be7-c594-4a72-aebd-0b746219231c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.281117] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.287270] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 77319462-e447-405b-9269-82581effe005 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.290778] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 757.290778] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 757.290778] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleting the datastore file [datastore2] efbde802-8cb0-4563-a776-3722a3889afe {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 757.291012] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d629eb3c-b9f4-462d-bbc3-88d948d9d392 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.297415] env[61768]: DEBUG oslo_vmware.api [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 757.297415] env[61768]: value = "task-1228764" [ 757.297415] env[61768]: _type = "Task" [ 757.297415] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.307787] env[61768]: DEBUG oslo_vmware.api [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228764, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.390923] env[61768]: DEBUG nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.462747] env[61768]: DEBUG nova.compute.manager [req-35598400-09af-4129-92a2-085ba838dff0 req-bb92b779-fc5f-4385-afba-1ec5e7af66f3 service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Received event network-vif-plugged-a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.462986] env[61768]: DEBUG oslo_concurrency.lockutils [req-35598400-09af-4129-92a2-085ba838dff0 req-bb92b779-fc5f-4385-afba-1ec5e7af66f3 service nova] Acquiring lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.463229] env[61768]: DEBUG oslo_concurrency.lockutils [req-35598400-09af-4129-92a2-085ba838dff0 req-bb92b779-fc5f-4385-afba-1ec5e7af66f3 service nova] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.463419] env[61768]: DEBUG oslo_concurrency.lockutils [req-35598400-09af-4129-92a2-085ba838dff0 req-bb92b779-fc5f-4385-afba-1ec5e7af66f3 service nova] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.463614] env[61768]: DEBUG nova.compute.manager [req-35598400-09af-4129-92a2-085ba838dff0 req-bb92b779-fc5f-4385-afba-1ec5e7af66f3 service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] No waiting events found dispatching network-vif-plugged-a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 757.463787] env[61768]: WARNING nova.compute.manager [req-35598400-09af-4129-92a2-085ba838dff0 req-bb92b779-fc5f-4385-afba-1ec5e7af66f3 service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Received unexpected event network-vif-plugged-a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 for instance with vm_state building and task_state spawning. [ 757.539543] env[61768]: DEBUG nova.network.neutron [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Successfully updated port: a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 757.582915] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228761, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.691686] env[61768]: DEBUG nova.network.neutron [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [{"id": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "address": "fa:16:3e:98:b7:f8", "network": {"id": "3e4a66df-17f2-479e-bb37-f8659f0180d6", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1225762599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ec5c061ca09040e8a2b3c3a6874daa40", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61e8c28f-c3", "ovs_interfaceid": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.726658] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228762, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.748639] env[61768]: DEBUG oslo_concurrency.lockutils [None req-55c32669-0cad-4d65-9b13-c20d6885b2ed tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.752067] env[61768]: DEBUG oslo_concurrency.lockutils [None req-55c32669-0cad-4d65-9b13-c20d6885b2ed tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 757.752067] env[61768]: DEBUG nova.objects.instance [None req-55c32669-0cad-4d65-9b13-c20d6885b2ed tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'flavor' on Instance uuid 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 757.790309] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance a428f5e0-ade1-4aa6-af9d-0e33efcfec62 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 757.810172] env[61768]: DEBUG oslo_vmware.api [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228764, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246643} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.813897] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 757.813897] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 757.813897] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 757.813897] env[61768]: INFO nova.compute.manager [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Took 1.12 seconds to destroy the instance on the hypervisor. [ 757.813897] env[61768]: DEBUG oslo.service.loopingcall [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.814216] env[61768]: DEBUG nova.compute.manager [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.814216] env[61768]: DEBUG nova.network.neutron [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 757.919825] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.042325] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "refresh_cache-659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.043103] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "refresh_cache-659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.043103] env[61768]: DEBUG nova.network.neutron [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.088346] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228761, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.194593] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Releasing lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.231490] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228762, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.253995] env[61768]: DEBUG nova.objects.instance [None req-55c32669-0cad-4d65-9b13-c20d6885b2ed tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'pci_requests' on Instance uuid 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 758.294758] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 62cddff5-d499-4ef3-869d-3cdb0328640d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.555174] env[61768]: DEBUG nova.network.neutron [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.584727] env[61768]: DEBUG oslo_vmware.api [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228761, 'name': PowerOnVM_Task, 'duration_secs': 1.022267} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.588113] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 758.588113] env[61768]: INFO nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Took 10.34 seconds to spawn the instance on the hypervisor. [ 758.588113] env[61768]: DEBUG nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 758.588113] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6688ca7-cad2-4878-be20-1a01f4fc1ea1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.596799] env[61768]: DEBUG nova.network.neutron [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 758.730662] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228762, 'name': Rename_Task, 'duration_secs': 1.152559} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.731272] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 758.731657] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-846765ea-f101-46dd-a2dc-18056a85f553 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.736604] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 758.736978] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ef13d8d-0170-4222-8c3c-e6e77cf20c0f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.740326] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 758.740326] env[61768]: value = "task-1228765" [ 758.740326] env[61768]: _type = "Task" [ 758.740326] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.744917] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 758.744917] env[61768]: value = "task-1228766" [ 758.744917] env[61768]: _type = "Task" [ 758.744917] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.753825] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.758018] env[61768]: DEBUG nova.objects.base [None req-55c32669-0cad-4d65-9b13-c20d6885b2ed tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Object Instance<12a543c2-1081-49d7-800b-07f0a2516904> lazy-loaded attributes: flavor,pci_requests {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 758.758018] env[61768]: DEBUG nova.network.neutron [None req-55c32669-0cad-4d65-9b13-c20d6885b2ed tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 758.761250] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228766, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.798114] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 2c85de5d-0111-40a6-a8bc-69c7eba0393d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 758.892976] env[61768]: DEBUG nova.network.neutron [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Updating instance_info_cache with network_info: [{"id": "a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0", "address": "fa:16:3e:85:4b:09", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5a53f41-9e", "ovs_interfaceid": "a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.935322] env[61768]: DEBUG oslo_concurrency.lockutils [None req-55c32669-0cad-4d65-9b13-c20d6885b2ed tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.184s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.059525] env[61768]: INFO nova.compute.manager [-] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Took 1.25 seconds to deallocate network for instance. [ 759.108966] env[61768]: INFO nova.compute.manager [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Took 30.26 seconds to build instance. [ 759.251768] env[61768]: DEBUG oslo_vmware.api [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228765, 'name': PowerOnVM_Task, 'duration_secs': 0.47206} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.252375] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 759.252611] env[61768]: INFO nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Took 8.51 seconds to spawn the instance on the hypervisor. [ 759.252805] env[61768]: DEBUG nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 759.253554] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c4e9243-b0a4-4802-9cab-292df5cec501 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.258733] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228766, 'name': PowerOffVM_Task, 'duration_secs': 0.296978} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.259263] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 759.259998] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefd0d22-1185-4c9b-96a7-d18e8def0071 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.281425] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962c18d5-633f-46c7-98ad-e76c39b2fe2b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.300745] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance fbb8c9db-e823-444a-84de-b231cf9b8a8a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.312279] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 759.312565] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a63d5dc-18ef-475b-9603-65e4a7389b20 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.319991] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 759.319991] env[61768]: value = "task-1228767" [ 759.319991] env[61768]: _type = "Task" [ 759.319991] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.328110] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228767, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.396044] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "refresh_cache-659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.396395] env[61768]: DEBUG nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Instance network_info: |[{"id": "a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0", "address": "fa:16:3e:85:4b:09", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5a53f41-9e", "ovs_interfaceid": "a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 759.396848] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:4b:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 759.405057] env[61768]: DEBUG oslo.service.loopingcall [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 759.405057] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 759.405057] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-238df0dd-a294-4b4b-8e67-693bf881cee6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.424546] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 759.424546] env[61768]: value = "task-1228768" [ 759.424546] env[61768]: _type = "Task" [ 759.424546] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.435153] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228768, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.564941] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.610236] env[61768]: DEBUG oslo_concurrency.lockutils [None req-67675bad-f676-48ee-bbb0-1407228ee453 tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.207s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.687878] env[61768]: DEBUG nova.compute.manager [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Received event network-changed-a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.688095] env[61768]: DEBUG nova.compute.manager [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Refreshing instance network info cache due to event network-changed-a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.688338] env[61768]: DEBUG oslo_concurrency.lockutils [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] Acquiring lock "refresh_cache-659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.688518] env[61768]: DEBUG oslo_concurrency.lockutils [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] Acquired lock "refresh_cache-659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.688691] env[61768]: DEBUG nova.network.neutron [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Refreshing network info cache for port a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 759.776235] env[61768]: INFO nova.compute.manager [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Took 28.49 seconds to build instance. [ 759.805344] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 3d96e47d-fc09-439e-bf0e-e60fc87ae958 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 759.805344] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 759.805344] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 759.826380] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 759.835482] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 759.836188] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 759.836188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.836188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.836336] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 759.837450] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7641d548-cfdf-4e82-b7f9-af9aa827d6c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.845150] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 759.845343] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 759.849412] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 759.849412] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 759.849861] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-419cd3d9-c965-4641-bc3e-f403af1849a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.855126] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 759.855126] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a97679-e4dd-bb33-c0b6-085a26e3c66b" [ 759.855126] env[61768]: _type = "Task" [ 759.855126] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.859050] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 759.865648] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a97679-e4dd-bb33-c0b6-085a26e3c66b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.879927] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 759.938012] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228768, 'name': CreateVM_Task, 'duration_secs': 0.368294} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.938487] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 759.939137] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.939354] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.939717] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 759.946944] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff57611c-8b9d-4cf4-a56b-b5b9c5224836 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.958026] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 759.958026] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1fba1-d70f-280f-5dac-72ee805922dd" [ 759.958026] env[61768]: _type = "Task" [ 759.958026] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.962751] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1fba1-d70f-280f-5dac-72ee805922dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.113480] env[61768]: DEBUG nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 760.180160] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9bee273-4521-42bb-b829-2f8328802ef2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.188239] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1926a0bd-3d3e-4c76-adab-28d806defdf4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.221609] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b5bc847-a45b-46b2-bc7b-2cf43587b9aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.231349] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7599eac2-4074-4e5c-b141-bec5ddd3a2e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.245460] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.279972] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edd3767c-0d2d-448e-b707-0fda77e4beb9 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.603s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.365868] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a97679-e4dd-bb33-c0b6-085a26e3c66b, 'name': SearchDatastore_Task, 'duration_secs': 0.038869} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.366700] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-938bcc2e-b26a-41d6-bd93-6ab35c75debc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.376024] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 760.376024] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524b65d9-3749-9fa4-0372-31b5eca9c443" [ 760.376024] env[61768]: _type = "Task" [ 760.376024] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.381574] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524b65d9-3749-9fa4-0372-31b5eca9c443, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.393625] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "d70040f1-e425-4aeb-b82d-ce808a02a645" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.393879] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.394157] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "d70040f1-e425-4aeb-b82d-ce808a02a645-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.394373] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.394558] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.397045] env[61768]: INFO nova.compute.manager [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Terminating instance [ 760.399173] env[61768]: DEBUG nova.compute.manager [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.399511] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 760.400353] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-108bd589-4dad-45e2-ac05-2eb3ad3696c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.408194] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 760.408457] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-75724c35-8063-4d6a-8da7-2f99f184981d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.414849] env[61768]: DEBUG oslo_vmware.api [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 760.414849] env[61768]: value = "task-1228769" [ 760.414849] env[61768]: _type = "Task" [ 760.414849] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.424163] env[61768]: DEBUG oslo_vmware.api [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.464949] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1fba1-d70f-280f-5dac-72ee805922dd, 'name': SearchDatastore_Task, 'duration_secs': 0.012157} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.465057] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.465310] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 760.465515] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.465673] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.465859] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 760.466479] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a1a95d31-f0a3-4223-b024-0406436f13c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.475524] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 760.475719] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 760.476743] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae731f5b-0e22-4329-a2dc-61412cbae5c7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.483137] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 760.483137] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52eea84a-72d7-0aa2-77fc-6df173e424e0" [ 760.483137] env[61768]: _type = "Task" [ 760.483137] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.486403] env[61768]: DEBUG nova.network.neutron [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Updated VIF entry in instance network info cache for port a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 760.486742] env[61768]: DEBUG nova.network.neutron [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Updating instance_info_cache with network_info: [{"id": "a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0", "address": "fa:16:3e:85:4b:09", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa5a53f41-9e", "ovs_interfaceid": "a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.493411] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52eea84a-72d7-0aa2-77fc-6df173e424e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.638332] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.749272] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.886900] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524b65d9-3749-9fa4-0372-31b5eca9c443, 'name': SearchDatastore_Task, 'duration_secs': 0.033323} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.887202] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.887500] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. {{(pid=61768) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 760.887782] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-639b1b5b-7eb7-47a0-bd4b-59a0d9ca3f0a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.897743] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 760.897743] env[61768]: value = "task-1228770" [ 760.897743] env[61768]: _type = "Task" [ 760.897743] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.912408] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228770, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.923831] env[61768]: DEBUG oslo_vmware.api [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228769, 'name': PowerOffVM_Task, 'duration_secs': 0.162906} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.924394] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 760.924616] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 760.924875] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d98de1ed-a03b-479c-99de-83fabb86a638 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.934821] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.935041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.935424] env[61768]: DEBUG nova.objects.instance [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'flavor' on Instance uuid 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 760.989404] env[61768]: DEBUG oslo_concurrency.lockutils [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] Releasing lock "refresh_cache-659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.989758] env[61768]: DEBUG nova.compute.manager [req-1cf3a3cf-7371-4c47-945f-5d275f2859a5 req-4e2768b7-9394-4a62-a4a5-5ec38dc0d2fc service nova] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Received event network-vif-deleted-5379c85e-3ce3-433a-ad59-3b790a40b5fb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 760.994511] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52eea84a-72d7-0aa2-77fc-6df173e424e0, 'name': SearchDatastore_Task, 'duration_secs': 0.013521} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.996304] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 760.996521] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 760.996795] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Deleting the datastore file [datastore2] d70040f1-e425-4aeb-b82d-ce808a02a645 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.997073] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c14054e3-31c2-4650-9350-7568ee5c8628 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.004048] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfe98c9c-70db-4ca4-9d40-8f3b851f76e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.007564] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 761.007564] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52818f56-15f3-5340-5b75-f25fb1352922" [ 761.007564] env[61768]: _type = "Task" [ 761.007564] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.009702] env[61768]: DEBUG oslo_vmware.api [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for the task: (returnval){ [ 761.009702] env[61768]: value = "task-1228772" [ 761.009702] env[61768]: _type = "Task" [ 761.009702] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.026096] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52818f56-15f3-5340-5b75-f25fb1352922, 'name': SearchDatastore_Task, 'duration_secs': 0.009691} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.029961] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.030335] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f/659e8d4c-1af6-4e81-b6a1-2d9466af6b0f.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 761.030669] env[61768]: DEBUG oslo_vmware.api [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228772, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.030898] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-86bbade6-c261-4e92-9671-8357f9e18b18 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.038097] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 761.038097] env[61768]: value = "task-1228773" [ 761.038097] env[61768]: _type = "Task" [ 761.038097] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.049125] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228773, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.254708] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 761.258029] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.178s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.258029] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.840s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.258029] env[61768]: INFO nova.compute.claims [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.261042] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 761.261278] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Cleaning up deleted instances {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 761.409074] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228770, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465979} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.409365] env[61768]: INFO nova.virt.vmwareapi.ds_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. [ 761.410228] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c69ac59-af70-4e40-a09a-273413acc4bf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.439885] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.443835] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-50de147e-b689-4eb8-a7ad-42237315ba20 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.463561] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 761.463561] env[61768]: value = "task-1228774" [ 761.463561] env[61768]: _type = "Task" [ 761.463561] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.473125] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228774, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.526340] env[61768]: DEBUG oslo_vmware.api [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Task: {'id': task-1228772, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.456584} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.526340] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.526340] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 761.526625] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 761.526728] env[61768]: INFO nova.compute.manager [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Took 1.13 seconds to destroy the instance on the hypervisor. [ 761.527066] env[61768]: DEBUG oslo.service.loopingcall [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.527339] env[61768]: DEBUG nova.compute.manager [-] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 761.527551] env[61768]: DEBUG nova.network.neutron [-] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 761.550728] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228773, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.643180] env[61768]: DEBUG nova.objects.instance [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'pci_requests' on Instance uuid 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 761.769513] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] There are 7 instances to clean {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 761.771101] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 3401be45-c3ab-4780-81a7-a93b05742414] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 761.976069] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228774, 'name': ReconfigVM_Task, 'duration_secs': 0.317573} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.976496] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.977669] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d79d09e-13ee-4b79-857a-72dbaf60cb48 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.007437] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-51beef7e-7202-4e55-9c50-22022a184ae3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.024527] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 762.024527] env[61768]: value = "task-1228775" [ 762.024527] env[61768]: _type = "Task" [ 762.024527] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.032909] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228775, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.048395] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228773, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.653098} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.048719] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f/659e8d4c-1af6-4e81-b6a1-2d9466af6b0f.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 762.048964] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.049274] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c324eb5e-6179-43d5-ad87-426408225a3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.056283] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 762.056283] env[61768]: value = "task-1228776" [ 762.056283] env[61768]: _type = "Task" [ 762.056283] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.064726] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228776, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.145910] env[61768]: DEBUG nova.objects.base [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Object Instance<12a543c2-1081-49d7-800b-07f0a2516904> lazy-loaded attributes: flavor,pci_requests {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 762.146103] env[61768]: DEBUG nova.network.neutron [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 762.194931] env[61768]: DEBUG nova.compute.manager [req-1858bdc6-2720-4bb6-9980-89ed51413dca req-f8e54c30-622a-4194-8ff4-394096498871 service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Received event network-vif-deleted-45566649-d3e3-41d9-9059-90cd8cb8411d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 762.195157] env[61768]: INFO nova.compute.manager [req-1858bdc6-2720-4bb6-9980-89ed51413dca req-f8e54c30-622a-4194-8ff4-394096498871 service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Neutron deleted interface 45566649-d3e3-41d9-9059-90cd8cb8411d; detaching it from the instance and deleting it from the info cache [ 762.195345] env[61768]: DEBUG nova.network.neutron [req-1858bdc6-2720-4bb6-9980-89ed51413dca req-f8e54c30-622a-4194-8ff4-394096498871 service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.206537] env[61768]: DEBUG nova.policy [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 762.272722] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 7a5f1b3f-847d-4f5c-8ac4-59f7c5e0a7ed] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 762.367099] env[61768]: DEBUG nova.network.neutron [-] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.536026] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228775, 'name': ReconfigVM_Task, 'duration_secs': 0.159395} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.536166] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 762.536458] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2e5f5f06-c20a-4a1d-aab8-ba21b32f8567 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.544917] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 762.544917] env[61768]: value = "task-1228777" [ 762.544917] env[61768]: _type = "Task" [ 762.544917] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.548626] env[61768]: DEBUG nova.network.neutron [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Successfully created port: a0589e32-427b-4fea-a7f6-6de811fcebdb {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.555723] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228777, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.567769] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228776, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065456} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.567769] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.567769] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3731ad91-345f-41c6-bf93-643812e4b1b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.590339] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f/659e8d4c-1af6-4e81-b6a1-2d9466af6b0f.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.592963] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f3ccd4a-f415-4051-b7fc-ccfe70d0fe79 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.615620] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 762.615620] env[61768]: value = "task-1228778" [ 762.615620] env[61768]: _type = "Task" [ 762.615620] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.622724] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228778, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.699273] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c05fc45-ed9f-493d-a831-dc15f8ff38c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.709070] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24a06785-1a60-4dbd-9f3b-0629eede10a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.737218] env[61768]: DEBUG nova.compute.manager [req-1858bdc6-2720-4bb6-9980-89ed51413dca req-f8e54c30-622a-4194-8ff4-394096498871 service nova] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Detach interface failed, port_id=45566649-d3e3-41d9-9059-90cd8cb8411d, reason: Instance d70040f1-e425-4aeb-b82d-ce808a02a645 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 762.742465] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b121a5f2-01c9-4030-ae15-0db907eb0999 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.751121] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa9efd7-1c0f-46d3-8e52-0b1a01345cb1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.785253] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 63704075-5eaf-4f84-a90b-3a0a3e904a9d] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 762.787948] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1e33ff-4c37-46f2-bb64-22b99e083cc5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.795934] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8bfef98-d8e3-4663-ad34-0981630742be {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.811170] env[61768]: DEBUG nova.compute.provider_tree [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.870535] env[61768]: INFO nova.compute.manager [-] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Took 1.34 seconds to deallocate network for instance. [ 763.056464] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228777, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.123734] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228778, 'name': ReconfigVM_Task, 'duration_secs': 0.29456} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.124047] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Reconfigured VM instance instance-0000003a to attach disk [datastore1] 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f/659e8d4c-1af6-4e81-b6a1-2d9466af6b0f.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.124688] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1fe302a2-3509-4028-8002-20fd70fd0a63 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.131065] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 763.131065] env[61768]: value = "task-1228779" [ 763.131065] env[61768]: _type = "Task" [ 763.131065] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.140666] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228779, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.289347] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 1eb76ef5-377c-4447-92da-4c61345c0070] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 763.315095] env[61768]: DEBUG nova.scheduler.client.report [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.379710] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.556374] env[61768]: DEBUG oslo_vmware.api [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228777, 'name': PowerOnVM_Task, 'duration_secs': 0.974799} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.556374] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 763.558690] env[61768]: DEBUG nova.compute.manager [None req-f44f6e95-a309-4c0d-a72a-91d160dbadb8 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 763.559538] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0abcbeb-3693-495b-894e-2a9d8a10146b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.644785] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228779, 'name': Rename_Task, 'duration_secs': 0.153973} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.644785] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 763.644785] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-844288f0-f0e6-490f-ab4a-ab37751fa887 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.650561] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 763.650561] env[61768]: value = "task-1228780" [ 763.650561] env[61768]: _type = "Task" [ 763.650561] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.661878] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228780, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.793197] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: f81aabf6-ae46-405c-b101-12ca707a0567] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 763.822147] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.822910] env[61768]: DEBUG nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.835315] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.442s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.837163] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.003s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.838499] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.943s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.838786] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.841851] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.490s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.842142] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.845020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.778s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.847895] env[61768]: INFO nova.compute.claims [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.893910] env[61768]: INFO nova.scheduler.client.report [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Deleted allocations for instance 91d56a8c-0664-4b3c-bf28-e2c668ac65fa [ 763.896218] env[61768]: INFO nova.scheduler.client.report [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted allocations for instance 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5 [ 763.914312] env[61768]: INFO nova.scheduler.client.report [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleted allocations for instance 0df161e9-4a66-432e-9366-60f231c53e3d [ 764.161814] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228780, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.255017] env[61768]: DEBUG nova.network.neutron [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Successfully updated port: a0589e32-427b-4fea-a7f6-6de811fcebdb {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 764.301047] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 0408fe1e-e4fa-4649-83dc-afdd335ee30d] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 764.335606] env[61768]: DEBUG nova.compute.utils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.339156] env[61768]: DEBUG nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 764.339365] env[61768]: DEBUG nova.network.neutron [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 764.403405] env[61768]: DEBUG nova.policy [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd1aa7c174648888b58f6a6df6f82a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfce0cf4a5d14042a929dac65aeb2ae3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 764.412118] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a704e05b-3847-4e60-9158-01f516cd1832 tempest-InstanceActionsTestJSON-1647655788 tempest-InstanceActionsTestJSON-1647655788-project-member] Lock "91d56a8c-0664-4b3c-bf28-e2c668ac65fa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.950s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.417867] env[61768]: DEBUG nova.compute.manager [req-294e354f-6cc6-4c19-b9d5-4a4e4f0d2406 req-153767ff-47b2-4a01-af74-022d34e532f1 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-vif-plugged-a0589e32-427b-4fea-a7f6-6de811fcebdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.418096] env[61768]: DEBUG oslo_concurrency.lockutils [req-294e354f-6cc6-4c19-b9d5-4a4e4f0d2406 req-153767ff-47b2-4a01-af74-022d34e532f1 service nova] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.418308] env[61768]: DEBUG oslo_concurrency.lockutils [req-294e354f-6cc6-4c19-b9d5-4a4e4f0d2406 req-153767ff-47b2-4a01-af74-022d34e532f1 service nova] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.418478] env[61768]: DEBUG oslo_concurrency.lockutils [req-294e354f-6cc6-4c19-b9d5-4a4e4f0d2406 req-153767ff-47b2-4a01-af74-022d34e532f1 service nova] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.418646] env[61768]: DEBUG nova.compute.manager [req-294e354f-6cc6-4c19-b9d5-4a4e4f0d2406 req-153767ff-47b2-4a01-af74-022d34e532f1 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] No waiting events found dispatching network-vif-plugged-a0589e32-427b-4fea-a7f6-6de811fcebdb {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 764.418836] env[61768]: WARNING nova.compute.manager [req-294e354f-6cc6-4c19-b9d5-4a4e4f0d2406 req-153767ff-47b2-4a01-af74-022d34e532f1 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received unexpected event network-vif-plugged-a0589e32-427b-4fea-a7f6-6de811fcebdb for instance with vm_state active and task_state None. [ 764.421016] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ce4aaa65-ae0b-4c9f-8f84-b0291a36e97b tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "663fa2fb-1ced-4052-9eba-8a5d0cbe96b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.298s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.426602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-80df0ace-7feb-49b5-82f2-5f39313913a5 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "0df161e9-4a66-432e-9366-60f231c53e3d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.131s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.661811] env[61768]: DEBUG oslo_vmware.api [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228780, 'name': PowerOnVM_Task, 'duration_secs': 0.527799} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.662108] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 764.662318] env[61768]: INFO nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Took 7.56 seconds to spawn the instance on the hypervisor. [ 764.662507] env[61768]: DEBUG nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 764.663285] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85dde2c5-351b-429d-a59d-25f983b1fde4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.760376] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.760376] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.760376] env[61768]: DEBUG nova.network.neutron [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.792206] env[61768]: DEBUG nova.network.neutron [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Successfully created port: b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 764.804438] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: ca465e62-c190-4d67-9dec-87ea1aa8f194] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 764.842397] env[61768]: DEBUG nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.162096] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0049623-f0b9-4260-a482-aec19ee75913 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.169525] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e28f9b7-a4ac-4fa6-82da-01ce2980a0ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.185429] env[61768]: INFO nova.compute.manager [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Took 31.91 seconds to build instance. [ 765.215945] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5cbf3b91-ec90-4c44-8ec8-4289ee78753c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.047s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.216776] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ea47e32-a428-4b1e-8018-2ba55d069836 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.228020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c4a846d-69a2-4276-85c8-1fd48231d6fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.241594] env[61768]: DEBUG nova.compute.provider_tree [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.310117] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.310351] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Cleaning up deleted instances with incomplete migration {{(pid=61768) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 765.333246] env[61768]: WARNING nova.network.neutron [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] 2fbe88f9-1bca-4347-914d-a62dc6bebde8 already exists in list: networks containing: ['2fbe88f9-1bca-4347-914d-a62dc6bebde8']. ignoring it [ 765.743981] env[61768]: DEBUG nova.scheduler.client.report [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.812988] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.851988] env[61768]: DEBUG nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.891547] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.891867] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.891966] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.893350] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.893542] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.893881] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.894438] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.894647] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.897341] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.897341] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.897341] env[61768]: DEBUG nova.virt.hardware [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.897341] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2c25cd5-8c3b-4ae3-a57f-c5360c4dda0b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.905867] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f05696b-54a1-42cb-a907-b7e05eebefa2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.008824] env[61768]: DEBUG nova.network.neutron [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "address": "fa:16:3e:7d:36:19", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0589e32-42", "ovs_interfaceid": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.249909] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.249909] env[61768]: DEBUG nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.253442] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.643s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.254832] env[61768]: INFO nova.compute.claims [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.516209] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.518085] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.518322] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.519251] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270da9f6-3982-41ca-936e-ecd8df00c255 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.550223] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.550573] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.550847] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.551512] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.551682] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.551902] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.552135] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.552388] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.552600] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.552886] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.553100] env[61768]: DEBUG nova.virt.hardware [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.562019] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfiguring VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 766.562019] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-953addcf-15ff-4026-8e1f-eb797dedbb67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.581192] env[61768]: DEBUG oslo_vmware.api [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 766.581192] env[61768]: value = "task-1228781" [ 766.581192] env[61768]: _type = "Task" [ 766.581192] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.588736] env[61768]: DEBUG oslo_vmware.api [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228781, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.764338] env[61768]: DEBUG nova.compute.utils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.771041] env[61768]: DEBUG nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.771264] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 766.901107] env[61768]: DEBUG nova.network.neutron [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Successfully updated port: b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 766.910636] env[61768]: DEBUG nova.policy [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5ed59d2bcee54afab032a65ae3b21023', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ecf64533220b4a5b9f05c0c864e408a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 767.092762] env[61768]: DEBUG oslo_vmware.api [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228781, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.275589] env[61768]: DEBUG nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.406204] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-76c282d1-ddfe-46dc-aa7b-225708443379" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.406361] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-76c282d1-ddfe-46dc-aa7b-225708443379" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.406514] env[61768]: DEBUG nova.network.neutron [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 767.563934] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Successfully created port: 5c1feaf9-e399-4171-90ae-e50db59e5274 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.597763] env[61768]: DEBUG oslo_vmware.api [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228781, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.655219] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc9587a-98d3-4156-9e7b-d28a692b3326 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.663868] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6622ccb7-c84b-4941-b016-88da88e527ca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.699937] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46c40f0-26d1-40c2-9d26-e7ba58db5e9c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.709253] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80405f7-a30a-4529-a107-821137967140 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.724706] env[61768]: DEBUG nova.compute.provider_tree [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.975547] env[61768]: DEBUG nova.network.neutron [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.979305] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Successfully created port: af1a7711-f21e-4462-93ef-a052655ce43c {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 768.094217] env[61768]: DEBUG oslo_vmware.api [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228781, 'name': ReconfigVM_Task, 'duration_secs': 1.081326} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.094748] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.095032] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfigured VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 768.228019] env[61768]: DEBUG nova.scheduler.client.report [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.243372] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.243741] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.285210] env[61768]: DEBUG nova.network.neutron [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Updating instance_info_cache with network_info: [{"id": "b4dfbd11-5359-4bf8-bd2f-f2704d5ead57", "address": "fa:16:3e:ac:74:e8", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4dfbd11-53", "ovs_interfaceid": "b4dfbd11-5359-4bf8-bd2f-f2704d5ead57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.290864] env[61768]: DEBUG nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.322213] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.322454] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.322611] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.322774] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.322930] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.323117] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.323343] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.323513] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.323691] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.323861] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.324061] env[61768]: DEBUG nova.virt.hardware [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.324928] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9efd0b64-5563-4a93-86c3-dc86aca571be {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.335347] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42d536d-00f3-471b-9ffd-9d78c0c93cf5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.522646] env[61768]: DEBUG nova.compute.manager [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-changed-a0589e32-427b-4fea-a7f6-6de811fcebdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.522898] env[61768]: DEBUG nova.compute.manager [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing instance network info cache due to event network-changed-a0589e32-427b-4fea-a7f6-6de811fcebdb. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.523375] env[61768]: DEBUG oslo_concurrency.lockutils [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.523634] env[61768]: DEBUG oslo_concurrency.lockutils [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.523919] env[61768]: DEBUG nova.network.neutron [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing network info cache for port a0589e32-427b-4fea-a7f6-6de811fcebdb {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 768.599314] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca84e681-3f73-437a-a45b-e5638b4d2d2f tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.664s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.736020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.736020] env[61768]: DEBUG nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.737271] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.481s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.739298] env[61768]: INFO nova.compute.claims [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.745729] env[61768]: DEBUG nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 768.789156] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-76c282d1-ddfe-46dc-aa7b-225708443379" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.789229] env[61768]: DEBUG nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Instance network_info: |[{"id": "b4dfbd11-5359-4bf8-bd2f-f2704d5ead57", "address": "fa:16:3e:ac:74:e8", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4dfbd11-53", "ovs_interfaceid": "b4dfbd11-5359-4bf8-bd2f-f2704d5ead57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 768.789632] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:74:e8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4dfbd11-5359-4bf8-bd2f-f2704d5ead57', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 768.798529] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating folder: Project (dfce0cf4a5d14042a929dac65aeb2ae3). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 768.799914] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7411d8d6-c52a-44af-be5b-0b53ebd9db5c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.814659] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created folder: Project (dfce0cf4a5d14042a929dac65aeb2ae3) in parent group-v265360. [ 768.814659] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating folder: Instances. Parent ref: group-v265424. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 768.814659] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c617f6c1-aec0-49c0-93cb-ff8c578eab08 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.822924] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created folder: Instances in parent group-v265424. [ 768.823192] env[61768]: DEBUG oslo.service.loopingcall [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 768.823389] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 768.823597] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4911a5a-8383-456d-be4a-74b11011a15d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.843936] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 768.843936] env[61768]: value = "task-1228784" [ 768.843936] env[61768]: _type = "Task" [ 768.843936] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.852254] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228784, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.245233] env[61768]: DEBUG nova.compute.utils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 769.250590] env[61768]: DEBUG nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 769.250770] env[61768]: DEBUG nova.network.neutron [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 769.281955] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.308261] env[61768]: DEBUG nova.policy [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '72ee09634b2f4644803e299e29032605', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c757f9fdbfc14f33948cc62c7ea93d53', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.356655] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228784, 'name': CreateVM_Task, 'duration_secs': 0.38817} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.356893] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 769.357698] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.358113] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.358212] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 769.358517] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6195c272-2084-4bf8-8c86-6c0b11c2f491 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.363859] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 769.363859] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f8c924-1dee-4ecf-e568-6817f01eefaf" [ 769.363859] env[61768]: _type = "Task" [ 769.363859] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.372261] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f8c924-1dee-4ecf-e568-6817f01eefaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.373328] env[61768]: DEBUG nova.network.neutron [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updated VIF entry in instance network info cache for port a0589e32-427b-4fea-a7f6-6de811fcebdb. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 769.373722] env[61768]: DEBUG nova.network.neutron [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "address": "fa:16:3e:7d:36:19", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0589e32-42", "ovs_interfaceid": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.628397] env[61768]: DEBUG nova.network.neutron [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Successfully created port: 0416fa92-b8f1-4664-b2e1-df607b4200a3 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.705797] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.706282] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.751574] env[61768]: DEBUG nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.885112] env[61768]: DEBUG oslo_concurrency.lockutils [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.885517] env[61768]: DEBUG nova.compute.manager [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 769.885785] env[61768]: DEBUG nova.compute.manager [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing instance network info cache due to event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 769.886113] env[61768]: DEBUG oslo_concurrency.lockutils [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] Acquiring lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.886337] env[61768]: DEBUG oslo_concurrency.lockutils [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] Acquired lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.886582] env[61768]: DEBUG nova.network.neutron [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.888169] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f8c924-1dee-4ecf-e568-6817f01eefaf, 'name': SearchDatastore_Task, 'duration_secs': 0.011535} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.888596] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.888867] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 769.889289] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.889523] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.889803] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 769.894275] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ed28c5ee-6076-4c0d-90d4-9d0cee3d35a8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.905538] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 769.909021] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 769.909021] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b63ee30a-a076-4b0f-9c5b-466e2651b57b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.914288] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 769.914288] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520f50f2-d3ad-2d57-e527-001d8fd14550" [ 769.914288] env[61768]: _type = "Task" [ 769.914288] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.927484] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520f50f2-d3ad-2d57-e527-001d8fd14550, 'name': SearchDatastore_Task, 'duration_secs': 0.009497} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.928476] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f511404b-4c44-426c-b20f-8eee3116c7b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.935674] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 769.935674] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cb71c4-5f11-7a77-c5b7-113566dd3a85" [ 769.935674] env[61768]: _type = "Task" [ 769.935674] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.945517] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cb71c4-5f11-7a77-c5b7-113566dd3a85, 'name': SearchDatastore_Task} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.945847] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.946070] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 76c282d1-ddfe-46dc-aa7b-225708443379/76c282d1-ddfe-46dc-aa7b-225708443379.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 769.946508] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d546c9f-4dc2-4d10-ac2e-2efa2b0d9a03 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.958200] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 769.958200] env[61768]: value = "task-1228785" [ 769.958200] env[61768]: _type = "Task" [ 769.958200] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.970356] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228785, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.116788] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Successfully updated port: 5c1feaf9-e399-4171-90ae-e50db59e5274 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 770.120247] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829f1103-b35c-4715-8b41-4f39af05c2bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.135088] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbe2742e-d92b-4455-a4a6-c90a43327836 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.172685] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8d7b61e-e1f5-4ec8-bf5c-00011c9cfefd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.186111] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edf0e522-5883-4f6b-baa2-12adcb4fb503 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.204100] env[61768]: DEBUG nova.compute.provider_tree [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.209208] env[61768]: DEBUG nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.469348] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228785, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447241} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.469668] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 76c282d1-ddfe-46dc-aa7b-225708443379/76c282d1-ddfe-46dc-aa7b-225708443379.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 770.469910] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 770.470295] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-acad4ef7-f7ad-4953-88c4-06632dc000b0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.478113] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 770.478113] env[61768]: value = "task-1228786" [ 770.478113] env[61768]: _type = "Task" [ 770.478113] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.486487] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228786, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.709146] env[61768]: DEBUG nova.scheduler.client.report [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.736868] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.763297] env[61768]: DEBUG nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.808879] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.809461] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.809461] env[61768]: DEBUG nova.virt.hardware [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.813368] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164c368e-dc58-4339-9904-09ba20090c69 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.821875] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d873ac5-485b-475b-af89-92aab7a27706 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.883909] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.884427] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.900459] env[61768]: DEBUG nova.network.neutron [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updated VIF entry in instance network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.901557] env[61768]: DEBUG nova.network.neutron [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [{"id": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "address": "fa:16:3e:98:b7:f8", "network": {"id": "3e4a66df-17f2-479e-bb37-f8659f0180d6", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1225762599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ec5c061ca09040e8a2b3c3a6874daa40", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61e8c28f-c3", "ovs_interfaceid": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.923288] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Received event network-vif-plugged-b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.923651] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquiring lock "76c282d1-ddfe-46dc-aa7b-225708443379-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.925917] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Lock "76c282d1-ddfe-46dc-aa7b-225708443379-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.925917] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Lock "76c282d1-ddfe-46dc-aa7b-225708443379-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.925917] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] No waiting events found dispatching network-vif-plugged-b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 770.925917] env[61768]: WARNING nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Received unexpected event network-vif-plugged-b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 for instance with vm_state building and task_state spawning. [ 770.925917] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Received event network-changed-b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.927057] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Refreshing instance network info cache due to event network-changed-b4dfbd11-5359-4bf8-bd2f-f2704d5ead57. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 770.927057] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquiring lock "refresh_cache-76c282d1-ddfe-46dc-aa7b-225708443379" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.927057] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquired lock "refresh_cache-76c282d1-ddfe-46dc-aa7b-225708443379" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.927057] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Refreshing network info cache for port b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 770.996118] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228786, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.210038} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.996451] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 770.997539] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb46fe87-0dd0-4777-b83e-ece06ba24146 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.028718] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 76c282d1-ddfe-46dc-aa7b-225708443379/76c282d1-ddfe-46dc-aa7b-225708443379.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.029107] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f820e56-c2d5-4cb9-a75d-13b92adc461c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.051355] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 771.051355] env[61768]: value = "task-1228787" [ 771.051355] env[61768]: _type = "Task" [ 771.051355] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.061754] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228787, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.212538] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.475s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.213196] env[61768]: DEBUG nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 771.216928] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.238s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.217224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.220155] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.211s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.223789] env[61768]: INFO nova.compute.claims [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.322571] env[61768]: INFO nova.scheduler.client.report [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Deleted allocations for instance 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3 [ 771.403934] env[61768]: DEBUG oslo_concurrency.lockutils [req-c1df5d3c-d96a-4387-b761-470151cd505c req-f23158e2-4515-4988-891a-2e4a7ba28f3a service nova] Releasing lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.467991] env[61768]: DEBUG nova.network.neutron [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Successfully updated port: 0416fa92-b8f1-4664-b2e1-df607b4200a3 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 771.562638] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228787, 'name': ReconfigVM_Task, 'duration_secs': 0.316498} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.562935] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 76c282d1-ddfe-46dc-aa7b-225708443379/76c282d1-ddfe-46dc-aa7b-225708443379.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 771.563658] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc81af5b-9933-47fb-b579-ae51806ce03d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.571270] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 771.571270] env[61768]: value = "task-1228788" [ 771.571270] env[61768]: _type = "Task" [ 771.571270] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.581099] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228788, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.728755] env[61768]: DEBUG nova.compute.utils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.731369] env[61768]: DEBUG nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.731577] env[61768]: DEBUG nova.network.neutron [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 771.770819] env[61768]: DEBUG nova.policy [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc4b77aedab34b96a303a133bc771e8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa61ecc36bc14769a2fc909ee95716ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.833272] env[61768]: DEBUG oslo_concurrency.lockutils [None req-956d87ea-12f8-4889-b468-3d5f23e8c676 tempest-ServerTagsTestJSON-60277261 tempest-ServerTagsTestJSON-60277261-project-member] Lock "7ef50ee4-64e6-49c2-bb41-0bdce316f8d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.903s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.913056] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Updated VIF entry in instance network info cache for port b4dfbd11-5359-4bf8-bd2f-f2704d5ead57. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 771.913467] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Updating instance_info_cache with network_info: [{"id": "b4dfbd11-5359-4bf8-bd2f-f2704d5ead57", "address": "fa:16:3e:ac:74:e8", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4dfbd11-53", "ovs_interfaceid": "b4dfbd11-5359-4bf8-bd2f-f2704d5ead57", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.970858] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "refresh_cache-a428f5e0-ade1-4aa6-af9d-0e33efcfec62" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.971180] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquired lock "refresh_cache-a428f5e0-ade1-4aa6-af9d-0e33efcfec62" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.971261] env[61768]: DEBUG nova.network.neutron [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 772.070608] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-cb8850cf-e632-419a-86f0-4c532e443213" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.071356] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-cb8850cf-e632-419a-86f0-4c532e443213" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.074019] env[61768]: DEBUG nova.objects.instance [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'flavor' on Instance uuid 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 772.087411] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228788, 'name': Rename_Task, 'duration_secs': 0.161957} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.087891] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 772.088299] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e7fffc9-d97a-47fa-a45e-b3290d4ba522 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.090821] env[61768]: DEBUG nova.network.neutron [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Successfully created port: 44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 772.099985] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 772.099985] env[61768]: value = "task-1228789" [ 772.099985] env[61768]: _type = "Task" [ 772.099985] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.112015] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.234021] env[61768]: DEBUG nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 772.416573] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Releasing lock "refresh_cache-76c282d1-ddfe-46dc-aa7b-225708443379" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.420980] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.420980] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing instance network info cache due to event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.420980] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquiring lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.420980] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquired lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.420980] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 772.509438] env[61768]: DEBUG nova.compute.manager [req-245d5b23-f008-44f6-8f04-dd72840a4a34 req-91beb3f9-2ae3-4c9c-9063-9f84d34e3b2f service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Received event network-vif-plugged-0416fa92-b8f1-4664-b2e1-df607b4200a3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.509650] env[61768]: DEBUG oslo_concurrency.lockutils [req-245d5b23-f008-44f6-8f04-dd72840a4a34 req-91beb3f9-2ae3-4c9c-9063-9f84d34e3b2f service nova] Acquiring lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.509892] env[61768]: DEBUG oslo_concurrency.lockutils [req-245d5b23-f008-44f6-8f04-dd72840a4a34 req-91beb3f9-2ae3-4c9c-9063-9f84d34e3b2f service nova] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.510440] env[61768]: DEBUG oslo_concurrency.lockutils [req-245d5b23-f008-44f6-8f04-dd72840a4a34 req-91beb3f9-2ae3-4c9c-9063-9f84d34e3b2f service nova] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.510716] env[61768]: DEBUG nova.compute.manager [req-245d5b23-f008-44f6-8f04-dd72840a4a34 req-91beb3f9-2ae3-4c9c-9063-9f84d34e3b2f service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] No waiting events found dispatching network-vif-plugged-0416fa92-b8f1-4664-b2e1-df607b4200a3 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 772.511028] env[61768]: WARNING nova.compute.manager [req-245d5b23-f008-44f6-8f04-dd72840a4a34 req-91beb3f9-2ae3-4c9c-9063-9f84d34e3b2f service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Received unexpected event network-vif-plugged-0416fa92-b8f1-4664-b2e1-df607b4200a3 for instance with vm_state building and task_state spawning. [ 772.518637] env[61768]: DEBUG nova.network.neutron [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.580474] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-774e0646-dfce-4131-bd98-54bbb4be6e79 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.592886] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1770a91e-d01c-460b-bc98-88a186ca642e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.628309] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6bbbff-a9b8-44e9-afde-e355fe907231 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.637203] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228789, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.641034] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9df01d5-4524-43d2-b89c-a04fecb8a898 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.658987] env[61768]: DEBUG nova.compute.provider_tree [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.867041] env[61768]: DEBUG nova.objects.instance [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'pci_requests' on Instance uuid 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 772.902576] env[61768]: DEBUG nova.network.neutron [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Updating instance_info_cache with network_info: [{"id": "0416fa92-b8f1-4664-b2e1-df607b4200a3", "address": "fa:16:3e:62:eb:64", "network": {"id": "33f1bcd3-7f8a-4a86-9e25-e1e14f3508d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1938220670-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c757f9fdbfc14f33948cc62c7ea93d53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cc448a80-6318-4b6a-b0a0-85fe6cc645df", "external-id": "nsx-vlan-transportzone-91", "segmentation_id": 91, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0416fa92-b8", "ovs_interfaceid": "0416fa92-b8f1-4664-b2e1-df607b4200a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.097865] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Successfully updated port: af1a7711-f21e-4462-93ef-a052655ce43c {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 773.138590] env[61768]: DEBUG oslo_vmware.api [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228789, 'name': PowerOnVM_Task, 'duration_secs': 0.640493} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.138948] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 773.139145] env[61768]: INFO nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Took 7.29 seconds to spawn the instance on the hypervisor. [ 773.139307] env[61768]: DEBUG nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 773.140524] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5040e18c-1b0b-465b-8491-f8e9fb651ced {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.162218] env[61768]: DEBUG nova.scheduler.client.report [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.247534] env[61768]: DEBUG nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 773.271476] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 773.271845] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 773.272109] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 773.272402] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 773.272662] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 773.272879] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 773.273226] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 773.273488] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 773.273762] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 773.274036] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 773.274315] env[61768]: DEBUG nova.virt.hardware [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 773.276228] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab72666d-825f-4901-8436-fe3813cfe59b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.287286] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345dae51-1d73-423e-b630-b0280876973f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.326395] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updated VIF entry in instance network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 773.326900] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [{"id": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "address": "fa:16:3e:98:b7:f8", "network": {"id": "3e4a66df-17f2-479e-bb37-f8659f0180d6", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1225762599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ec5c061ca09040e8a2b3c3a6874daa40", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61e8c28f-c3", "ovs_interfaceid": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.372875] env[61768]: DEBUG nova.objects.base [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Object Instance<12a543c2-1081-49d7-800b-07f0a2516904> lazy-loaded attributes: flavor,pci_requests {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 773.373126] env[61768]: DEBUG nova.network.neutron [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 773.405081] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Releasing lock "refresh_cache-a428f5e0-ade1-4aa6-af9d-0e33efcfec62" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.405394] env[61768]: DEBUG nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Instance network_info: |[{"id": "0416fa92-b8f1-4664-b2e1-df607b4200a3", "address": "fa:16:3e:62:eb:64", "network": {"id": "33f1bcd3-7f8a-4a86-9e25-e1e14f3508d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1938220670-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c757f9fdbfc14f33948cc62c7ea93d53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cc448a80-6318-4b6a-b0a0-85fe6cc645df", "external-id": "nsx-vlan-transportzone-91", "segmentation_id": 91, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0416fa92-b8", "ovs_interfaceid": "0416fa92-b8f1-4664-b2e1-df607b4200a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 773.405907] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:62:eb:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cc448a80-6318-4b6a-b0a0-85fe6cc645df', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0416fa92-b8f1-4664-b2e1-df607b4200a3', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 773.413531] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Creating folder: Project (c757f9fdbfc14f33948cc62c7ea93d53). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.413828] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cee357b6-848c-4edc-a904-c03adba6e29c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.427958] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Created folder: Project (c757f9fdbfc14f33948cc62c7ea93d53) in parent group-v265360. [ 773.428173] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Creating folder: Instances. Parent ref: group-v265427. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 773.428435] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcae5f07-5a98-4c99-86cd-1f53ff927787 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.439313] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Created folder: Instances in parent group-v265427. [ 773.439511] env[61768]: DEBUG oslo.service.loopingcall [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.439717] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 773.439928] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e8fb486c-464a-4b31-8954-76e0e94362c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.456581] env[61768]: DEBUG nova.policy [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 773.464456] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 773.464456] env[61768]: value = "task-1228792" [ 773.464456] env[61768]: _type = "Task" [ 773.464456] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.478395] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228792, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.600326] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.600601] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquired lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.600649] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 773.628968] env[61768]: DEBUG nova.network.neutron [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Successfully updated port: 44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 773.659535] env[61768]: INFO nova.compute.manager [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Took 39.27 seconds to build instance. [ 773.671408] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.671951] env[61768]: DEBUG nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 773.675935] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.888s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.677808] env[61768]: INFO nova.compute.claims [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.830500] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Releasing lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.830500] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received event network-vif-plugged-5c1feaf9-e399-4171-90ae-e50db59e5274 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.830500] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquiring lock "77319462-e447-405b-9269-82581effe005-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.830826] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Lock "77319462-e447-405b-9269-82581effe005-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.830826] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Lock "77319462-e447-405b-9269-82581effe005-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.834029] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] No waiting events found dispatching network-vif-plugged-5c1feaf9-e399-4171-90ae-e50db59e5274 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 773.834029] env[61768]: WARNING nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received unexpected event network-vif-plugged-5c1feaf9-e399-4171-90ae-e50db59e5274 for instance with vm_state building and task_state spawning. [ 773.834029] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received event network-changed-5c1feaf9-e399-4171-90ae-e50db59e5274 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.834029] env[61768]: DEBUG nova.compute.manager [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] Refreshing instance network info cache due to event network-changed-5c1feaf9-e399-4171-90ae-e50db59e5274. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 773.834029] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquiring lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.977255] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228792, 'name': CreateVM_Task, 'duration_secs': 0.40596} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.977513] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 773.978313] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.978536] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.979019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 773.979677] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9f8c190-7520-424a-8bea-cf81e3530e4f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.990745] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 773.990745] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523fac5f-2890-b00d-a14e-00851f3e5ec7" [ 773.990745] env[61768]: _type = "Task" [ 773.990745] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.000961] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523fac5f-2890-b00d-a14e-00851f3e5ec7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.122649] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "d928f14c-b94a-4abf-a053-51015f3bc6c5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.122894] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.131577] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.131721] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.131872] env[61768]: DEBUG nova.network.neutron [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 774.144759] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.152899] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "356a69b0-7898-440d-9473-7b4572ed2315" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.153159] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "356a69b0-7898-440d-9473-7b4572ed2315" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.162570] env[61768]: DEBUG oslo_concurrency.lockutils [None req-513b558e-14e2-4da8-b875-4f197fe9ab15 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "76c282d1-ddfe-46dc-aa7b-225708443379" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.992s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.179942] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "b6e4f743-5692-4b46-892c-6c8917ccef98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.179942] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.182255] env[61768]: DEBUG nova.compute.utils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 774.189065] env[61768]: DEBUG nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 774.189065] env[61768]: DEBUG nova.network.neutron [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 774.282520] env[61768]: DEBUG nova.policy [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fd5852bc16429aa2904b4351c8249d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7857b5217cc499ca8c9c3b8a5471db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 774.431339] env[61768]: DEBUG nova.network.neutron [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Updating instance_info_cache with network_info: [{"id": "5c1feaf9-e399-4171-90ae-e50db59e5274", "address": "fa:16:3e:90:0b:b0", "network": {"id": "936f4d95-f203-43f8-ba39-ef22c1c8ea4e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1357034332", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.83", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c1feaf9-e3", "ovs_interfaceid": "5c1feaf9-e399-4171-90ae-e50db59e5274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "af1a7711-f21e-4462-93ef-a052655ce43c", "address": "fa:16:3e:e3:57:9b", "network": {"id": "d3307f81-acee-4479-a53a-a0c40369ec9f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1851008898", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf1a7711-f2", "ovs_interfaceid": "af1a7711-f21e-4462-93ef-a052655ce43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.507805] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523fac5f-2890-b00d-a14e-00851f3e5ec7, 'name': SearchDatastore_Task, 'duration_secs': 0.01462} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.508146] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.508395] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 774.508657] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.508833] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.509041] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 774.510339] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7df0ba0a-0e0a-423f-9d5b-38f2a86c3f2e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.519422] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 774.519614] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 774.520356] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-664dd84c-401e-4fb9-8566-8160301de2e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.527249] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 774.527249] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52767ffe-3f66-597a-201a-20d44fc7c3c0" [ 774.527249] env[61768]: _type = "Task" [ 774.527249] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.536440] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52767ffe-3f66-597a-201a-20d44fc7c3c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.663346] env[61768]: DEBUG nova.network.neutron [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.665405] env[61768]: DEBUG nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 774.691959] env[61768]: DEBUG nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 774.846010] env[61768]: DEBUG nova.network.neutron [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.918776] env[61768]: DEBUG nova.network.neutron [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Successfully created port: 49a82793-2751-49db-b262-593b386ea68d {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 774.934180] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Releasing lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.934551] env[61768]: DEBUG nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Instance network_info: |[{"id": "5c1feaf9-e399-4171-90ae-e50db59e5274", "address": "fa:16:3e:90:0b:b0", "network": {"id": "936f4d95-f203-43f8-ba39-ef22c1c8ea4e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1357034332", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.83", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c1feaf9-e3", "ovs_interfaceid": "5c1feaf9-e399-4171-90ae-e50db59e5274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "af1a7711-f21e-4462-93ef-a052655ce43c", "address": "fa:16:3e:e3:57:9b", "network": {"id": "d3307f81-acee-4479-a53a-a0c40369ec9f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1851008898", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf1a7711-f2", "ovs_interfaceid": "af1a7711-f21e-4462-93ef-a052655ce43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 774.934873] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Acquired lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.935076] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] Refreshing network info cache for port 5c1feaf9-e399-4171-90ae-e50db59e5274 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 774.936260] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:0b:b0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '60badc2d-69d2-467d-a92e-98511f5cb0b2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5c1feaf9-e399-4171-90ae-e50db59e5274', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:e3:57:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c4349e30-c086-4c24-9e0e-83996d808a1b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'af1a7711-f21e-4462-93ef-a052655ce43c', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 774.947645] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Creating folder: Project (ecf64533220b4a5b9f05c0c864e408a0). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 774.951489] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4094f242-372e-4bde-af1e-cc38b2978fb0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.965570] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Created folder: Project (ecf64533220b4a5b9f05c0c864e408a0) in parent group-v265360. [ 774.965910] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Creating folder: Instances. Parent ref: group-v265430. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 774.969759] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d5b2763e-306f-4232-b63f-0d11bd9d7f30 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.982009] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Created folder: Instances in parent group-v265430. [ 774.982322] env[61768]: DEBUG oslo.service.loopingcall [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.982539] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77319462-e447-405b-9269-82581effe005] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 774.982798] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74aef02b-96c4-4699-b229-c07bb19d7428 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.014536] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.014536] env[61768]: value = "task-1228795" [ 775.014536] env[61768]: _type = "Task" [ 775.014536] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.024381] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228795, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.036677] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52767ffe-3f66-597a-201a-20d44fc7c3c0, 'name': SearchDatastore_Task, 'duration_secs': 0.013759} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.038041] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-75a1d80d-876c-4644-a4a9-37ba71fa2000 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.052215] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 775.052215] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52117c6e-46d7-f262-3ff1-4ba6c49b6c45" [ 775.052215] env[61768]: _type = "Task" [ 775.052215] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.063888] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52117c6e-46d7-f262-3ff1-4ba6c49b6c45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.107233] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33ed1192-bbdd-418f-b59e-4b9785c269e7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.114626] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-411ec3a7-77ca-4c07-85bd-f5fd87fd6d83 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.150034] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9decb83d-b731-4807-b630-502a3040180b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.160988] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c0470b-c3fa-4ef5-8db8-31b6e61e5719 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.181645] env[61768]: DEBUG nova.compute.provider_tree [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 775.201296] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.348971] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.349349] env[61768]: DEBUG nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance network_info: |[{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 775.349919] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:98:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37434b93-dfdc-4a3f-bf5a-9f2cbe25a754', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44b2acbd-7a9f-416b-b3b4-e877de5bfa85', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 775.357796] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating folder: Project (aa61ecc36bc14769a2fc909ee95716ae). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.358264] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8768fa0b-eb39-4945-af41-f3e762285cb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.373388] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created folder: Project (aa61ecc36bc14769a2fc909ee95716ae) in parent group-v265360. [ 775.373654] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating folder: Instances. Parent ref: group-v265433. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 775.373964] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28d75dc0-588f-466c-a922-ea9e52605fba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.388060] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created folder: Instances in parent group-v265433. [ 775.388356] env[61768]: DEBUG oslo.service.loopingcall [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.388578] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 775.388838] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4a473ba-1f00-4a8e-aabf-e74f8e7d1f36 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.409781] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 775.409781] env[61768]: value = "task-1228798" [ 775.409781] env[61768]: _type = "Task" [ 775.409781] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.419168] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228798, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.525291] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228795, 'name': CreateVM_Task, 'duration_secs': 0.459659} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.527623] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77319462-e447-405b-9269-82581effe005] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 775.528429] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 775.528634] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 775.528966] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 775.529535] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbc4c013-97bb-45a2-b1ed-3ad7d573a1b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.534939] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 775.534939] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5243f36d-3ee4-7057-48f1-d778149a6322" [ 775.534939] env[61768]: _type = "Task" [ 775.534939] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.543536] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5243f36d-3ee4-7057-48f1-d778149a6322, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.563564] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52117c6e-46d7-f262-3ff1-4ba6c49b6c45, 'name': SearchDatastore_Task, 'duration_secs': 0.012093} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.563846] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.564136] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] a428f5e0-ade1-4aa6-af9d-0e33efcfec62/a428f5e0-ade1-4aa6-af9d-0e33efcfec62.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 775.564415] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff4984bc-cc75-4f73-8fcf-72cb1733382d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.573518] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 775.573518] env[61768]: value = "task-1228799" [ 775.573518] env[61768]: _type = "Task" [ 775.573518] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.584794] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228799, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.689322] env[61768]: DEBUG nova.scheduler.client.report [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.702771] env[61768]: DEBUG nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 775.743153] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 775.743548] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 775.743738] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 775.743938] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 775.744126] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 775.744344] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 775.744632] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 775.744834] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 775.745064] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 775.745273] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 775.745595] env[61768]: DEBUG nova.virt.hardware [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 775.747368] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-270fbc9e-dddd-4377-9710-c0eeb36b280e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.759967] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f7f598-641c-41d5-b632-222788788882 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.892797] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] Updated VIF entry in instance network info cache for port 5c1feaf9-e399-4171-90ae-e50db59e5274. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 775.893424] env[61768]: DEBUG nova.network.neutron [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] [instance: 77319462-e447-405b-9269-82581effe005] Updating instance_info_cache with network_info: [{"id": "5c1feaf9-e399-4171-90ae-e50db59e5274", "address": "fa:16:3e:90:0b:b0", "network": {"id": "936f4d95-f203-43f8-ba39-ef22c1c8ea4e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1357034332", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.83", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c1feaf9-e3", "ovs_interfaceid": "5c1feaf9-e399-4171-90ae-e50db59e5274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "af1a7711-f21e-4462-93ef-a052655ce43c", "address": "fa:16:3e:e3:57:9b", "network": {"id": "d3307f81-acee-4479-a53a-a0c40369ec9f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1851008898", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf1a7711-f2", "ovs_interfaceid": "af1a7711-f21e-4462-93ef-a052655ce43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.927453] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228798, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.046378] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5243f36d-3ee4-7057-48f1-d778149a6322, 'name': SearchDatastore_Task, 'duration_secs': 0.010629} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.046712] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.046954] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.047279] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.047392] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.047581] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 776.048190] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-779721ad-2a50-4538-96df-ff85c90ef2fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.058302] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 776.058544] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 776.059292] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cd273cd-c4df-40ee-b361-ad4505618e89 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.065619] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 776.065619] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526152d5-3ba9-2190-f435-2f1bb1380679" [ 776.065619] env[61768]: _type = "Task" [ 776.065619] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.074831] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526152d5-3ba9-2190-f435-2f1bb1380679, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.083345] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228799, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.196983] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.521s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.197602] env[61768]: DEBUG nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 776.200605] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.912s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.200956] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.207819] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.883s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.207819] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.209357] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.216s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.210494] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.212664] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.634s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.212994] env[61768]: DEBUG nova.objects.instance [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 776.245600] env[61768]: DEBUG nova.network.neutron [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Successfully updated port: cb8850cf-e632-419a-86f0-4c532e443213 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.249860] env[61768]: INFO nova.scheduler.client.report [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted allocations for instance 9fa32a4d-3ba4-4d36-963b-17a64453e804 [ 776.257128] env[61768]: INFO nova.scheduler.client.report [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Deleted allocations for instance 9b3632e5-4c2d-4968-9692-70f91883bfb3 [ 776.279726] env[61768]: INFO nova.scheduler.client.report [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted allocations for instance 668d4fe4-85f8-4282-8d65-0549d9e3bda8 [ 776.396375] env[61768]: DEBUG oslo_concurrency.lockutils [req-fce6cd2c-8e72-48af-b144-a5a83b23a716 req-f4b3ea3b-828b-42e3-bdc5-63e003d5afca service nova] Releasing lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.427242] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228798, 'name': CreateVM_Task, 'duration_secs': 0.587196} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.427447] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 776.428224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.428412] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.428749] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 776.429030] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-716916e6-1dfd-427f-9255-94290753a2b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.435424] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 776.435424] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52281a8f-9e4c-5ff3-8b33-12b0ea25dced" [ 776.435424] env[61768]: _type = "Task" [ 776.435424] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.444382] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52281a8f-9e4c-5ff3-8b33-12b0ea25dced, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.581601] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526152d5-3ba9-2190-f435-2f1bb1380679, 'name': SearchDatastore_Task, 'duration_secs': 0.027177} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.587128] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed2a6683-01cf-4b65-ba78-a0bca7f945f7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.595158] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 776.595158] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e47655-7662-57f2-b3eb-8230cd34dcfa" [ 776.595158] env[61768]: _type = "Task" [ 776.595158] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.599235] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228799, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.535507} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.602457] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] a428f5e0-ade1-4aa6-af9d-0e33efcfec62/a428f5e0-ade1-4aa6-af9d-0e33efcfec62.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 776.602706] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 776.603067] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-20e2b000-780a-49a6-815d-d31e8054eba1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.612481] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e47655-7662-57f2-b3eb-8230cd34dcfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.614340] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 776.614340] env[61768]: value = "task-1228800" [ 776.614340] env[61768]: _type = "Task" [ 776.614340] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.623897] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228800, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.719959] env[61768]: DEBUG nova.compute.utils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.725620] env[61768]: DEBUG nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 776.726046] env[61768]: DEBUG nova.network.neutron [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 776.749797] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.749797] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.749797] env[61768]: DEBUG nova.network.neutron [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.768927] env[61768]: DEBUG oslo_concurrency.lockutils [None req-121774d5-e94c-4f4e-901a-fa94b6c96889 tempest-ServerAddressesNegativeTestJSON-20234865 tempest-ServerAddressesNegativeTestJSON-20234865-project-member] Lock "9b3632e5-4c2d-4968-9692-70f91883bfb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.473s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.770605] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8328c956-a051-4562-9529-2d9527725fa5 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "9fa32a4d-3ba4-4d36-963b-17a64453e804" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.610s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.793678] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c48a9253-785e-44bb-a469-00c659203d64 tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "668d4fe4-85f8-4282-8d65-0549d9e3bda8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.715s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.833973] env[61768]: DEBUG nova.policy [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7b1a8a74422b48e78a177a4b0a0709b0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f595b30e898b426a9675202e54338a4b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.865905] env[61768]: DEBUG nova.compute.manager [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 776.866218] env[61768]: DEBUG nova.compute.manager [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing instance network info cache due to event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 776.866397] env[61768]: DEBUG oslo_concurrency.lockutils [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] Acquiring lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.866557] env[61768]: DEBUG oslo_concurrency.lockutils [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] Acquired lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.866711] env[61768]: DEBUG nova.network.neutron [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 776.947366] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52281a8f-9e4c-5ff3-8b33-12b0ea25dced, 'name': SearchDatastore_Task, 'duration_secs': 0.031683} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.947784] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 776.948091] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 776.948333] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.114584] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e47655-7662-57f2-b3eb-8230cd34dcfa, 'name': SearchDatastore_Task, 'duration_secs': 0.017639} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.115160] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.115727] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 77319462-e447-405b-9269-82581effe005/77319462-e447-405b-9269-82581effe005.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 777.120773] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.121131] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 777.121618] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5770b0b-26eb-4d44-a4f0-bd1004654fcf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.125503] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e80549de-2660-4e65-9889-2611e970364f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.139206] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228800, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.253904} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.141069] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 777.141959] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 777.142566] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 777.143134] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 777.143134] env[61768]: value = "task-1228801" [ 777.143134] env[61768]: _type = "Task" [ 777.143134] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.144335] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57e54ac-eb1a-4c4a-81dd-bc3022be4305 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.147099] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-105116ac-295b-4e3c-a72d-65c54137afb8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.160731] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 777.160731] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5254a9c8-7541-0b30-4c7a-197fe1251554" [ 777.160731] env[61768]: _type = "Task" [ 777.160731] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.161713] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228801, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.182160] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] a428f5e0-ade1-4aa6-af9d-0e33efcfec62/a428f5e0-ade1-4aa6-af9d-0e33efcfec62.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 777.186824] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-731fcee8-a388-471d-ba21-91757f90ebf9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.204333] env[61768]: DEBUG nova.compute.manager [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Received event network-changed-0416fa92-b8f1-4664-b2e1-df607b4200a3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 777.204531] env[61768]: DEBUG nova.compute.manager [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Refreshing instance network info cache due to event network-changed-0416fa92-b8f1-4664-b2e1-df607b4200a3. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 777.204754] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Acquiring lock "refresh_cache-a428f5e0-ade1-4aa6-af9d-0e33efcfec62" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.204907] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Acquired lock "refresh_cache-a428f5e0-ade1-4aa6-af9d-0e33efcfec62" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.205182] env[61768]: DEBUG nova.network.neutron [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Refreshing network info cache for port 0416fa92-b8f1-4664-b2e1-df607b4200a3 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 777.213991] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5254a9c8-7541-0b30-4c7a-197fe1251554, 'name': SearchDatastore_Task, 'duration_secs': 0.047532} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.216471] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 777.216471] env[61768]: value = "task-1228802" [ 777.216471] env[61768]: _type = "Task" [ 777.216471] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.216888] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-438c083a-0205-4b5c-8b33-90c9803b750e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.226294] env[61768]: DEBUG nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 777.236119] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7f50d590-0c6c-4d78-9250-b23098b9c082 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.022s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.236206] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 777.236206] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524759dc-6035-ad3d-4ab7-dd99554e49c7" [ 777.236206] env[61768]: _type = "Task" [ 777.236206] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.236431] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228802, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.236693] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.317s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.238662] env[61768]: INFO nova.compute.claims [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 777.257256] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524759dc-6035-ad3d-4ab7-dd99554e49c7, 'name': SearchDatastore_Task, 'duration_secs': 0.023419} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.257593] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.257841] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 777.258446] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b100c104-27fe-4824-afb5-f75b2c01960e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.267973] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 777.267973] env[61768]: value = "task-1228803" [ 777.267973] env[61768]: _type = "Task" [ 777.267973] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.277310] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228803, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.296848] env[61768]: DEBUG nova.network.neutron [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Successfully created port: 3bc689b6-7278-4741-8053-b98ed2db0265 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 777.322582] env[61768]: WARNING nova.network.neutron [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] 2fbe88f9-1bca-4347-914d-a62dc6bebde8 already exists in list: networks containing: ['2fbe88f9-1bca-4347-914d-a62dc6bebde8']. ignoring it [ 777.322867] env[61768]: WARNING nova.network.neutron [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] 2fbe88f9-1bca-4347-914d-a62dc6bebde8 already exists in list: networks containing: ['2fbe88f9-1bca-4347-914d-a62dc6bebde8']. ignoring it [ 777.661535] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228801, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.746769] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228802, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.787175] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228803, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.019138] env[61768]: DEBUG nova.network.neutron [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Successfully updated port: 49a82793-2751-49db-b262-593b386ea68d {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 778.073865] env[61768]: DEBUG nova.network.neutron [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updated VIF entry in instance network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 778.074257] env[61768]: DEBUG nova.network.neutron [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [{"id": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "address": "fa:16:3e:98:b7:f8", "network": {"id": "3e4a66df-17f2-479e-bb37-f8659f0180d6", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1225762599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ec5c061ca09040e8a2b3c3a6874daa40", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61e8c28f-c3", "ovs_interfaceid": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.158406] env[61768]: DEBUG nova.network.neutron [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "address": "fa:16:3e:7d:36:19", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0589e32-42", "ovs_interfaceid": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cb8850cf-e632-419a-86f0-4c532e443213", "address": "fa:16:3e:c8:6d:50", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb8850cf-e6", "ovs_interfaceid": "cb8850cf-e632-419a-86f0-4c532e443213", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.169230] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228801, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.97919} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.169601] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 77319462-e447-405b-9269-82581effe005/77319462-e447-405b-9269-82581effe005.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 778.171129] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.171129] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-af41fd62-0479-4b93-8e3f-af228f148e68 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.181696] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 778.181696] env[61768]: value = "task-1228804" [ 778.181696] env[61768]: _type = "Task" [ 778.181696] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.196572] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228804, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.237363] env[61768]: DEBUG nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.240419] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228802, 'name': ReconfigVM_Task, 'duration_secs': 0.742867} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.241670] env[61768]: DEBUG nova.network.neutron [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Updated VIF entry in instance network info cache for port 0416fa92-b8f1-4664-b2e1-df607b4200a3. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 778.242176] env[61768]: DEBUG nova.network.neutron [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Updating instance_info_cache with network_info: [{"id": "0416fa92-b8f1-4664-b2e1-df607b4200a3", "address": "fa:16:3e:62:eb:64", "network": {"id": "33f1bcd3-7f8a-4a86-9e25-e1e14f3508d8", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1938220670-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c757f9fdbfc14f33948cc62c7ea93d53", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cc448a80-6318-4b6a-b0a0-85fe6cc645df", "external-id": "nsx-vlan-transportzone-91", "segmentation_id": 91, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0416fa92-b8", "ovs_interfaceid": "0416fa92-b8f1-4664-b2e1-df607b4200a3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.244807] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Reconfigured VM instance instance-0000003d to attach disk [datastore1] a428f5e0-ade1-4aa6-af9d-0e33efcfec62/a428f5e0-ade1-4aa6-af9d-0e33efcfec62.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 778.246488] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67a21d6e-77f7-4b41-87b2-07c02dfd197e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.261337] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 778.261337] env[61768]: value = "task-1228805" [ 778.261337] env[61768]: _type = "Task" [ 778.261337] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.280906] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228805, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.287979] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.288454] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.289211] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.289211] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.289460] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.289720] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.290069] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.290334] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.290635] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.290892] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.291179] env[61768]: DEBUG nova.virt.hardware [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.292962] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09e91d0f-1c76-4730-aa65-7b7903405bb2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.299650] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228803, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.305593] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9f6882-cbee-43af-99a5-98237021e02a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.522855] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.523132] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.523367] env[61768]: DEBUG nova.network.neutron [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.577732] env[61768]: DEBUG oslo_concurrency.lockutils [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] Releasing lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.578021] env[61768]: DEBUG nova.compute.manager [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received event network-vif-plugged-af1a7711-f21e-4462-93ef-a052655ce43c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.578245] env[61768]: DEBUG oslo_concurrency.lockutils [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] Acquiring lock "77319462-e447-405b-9269-82581effe005-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.578453] env[61768]: DEBUG oslo_concurrency.lockutils [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] Lock "77319462-e447-405b-9269-82581effe005-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.578618] env[61768]: DEBUG oslo_concurrency.lockutils [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] Lock "77319462-e447-405b-9269-82581effe005-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.578790] env[61768]: DEBUG nova.compute.manager [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 77319462-e447-405b-9269-82581effe005] No waiting events found dispatching network-vif-plugged-af1a7711-f21e-4462-93ef-a052655ce43c {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 778.578957] env[61768]: WARNING nova.compute.manager [req-0452a17a-711e-4aa2-8026-015b2136635f req-3ad568e6-15f5-478b-8208-2764d5002422 service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received unexpected event network-vif-plugged-af1a7711-f21e-4462-93ef-a052655ce43c for instance with vm_state building and task_state spawning. [ 778.581590] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82f7529-e6e5-4826-be3e-7f1052250495 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.589232] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb7d455-7e5e-4088-b0e4-1bd2f751e3ca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.619421] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f5f656-6e9e-4305-bc95-345705518635 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.627780] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75785b10-1c23-4f53-8e98-afafe0385157 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.643098] env[61768]: DEBUG nova.compute.provider_tree [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.661392] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.662080] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.662243] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.663444] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66afa025-b67c-4efa-9c25-8d22a9037e65 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.682177] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.686018] env[61768]: DEBUG nova.virt.hardware [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.690859] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfiguring VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 778.691287] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-159e0f27-2ab5-4586-bd68-a9ed7955b657 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.714456] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228804, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.183835} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.716036] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 778.716436] env[61768]: DEBUG oslo_vmware.api [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 778.716436] env[61768]: value = "task-1228806" [ 778.716436] env[61768]: _type = "Task" [ 778.716436] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.717145] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b996f50-5b56-451e-ba67-b3136a19c122 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.729130] env[61768]: DEBUG oslo_vmware.api [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228806, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.750838] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 77319462-e447-405b-9269-82581effe005/77319462-e447-405b-9269-82581effe005.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 778.751505] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Releasing lock "refresh_cache-a428f5e0-ade1-4aa6-af9d-0e33efcfec62" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.751752] env[61768]: DEBUG nova.compute.manager [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-vif-plugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.751954] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.752234] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.752508] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.752770] env[61768]: DEBUG nova.compute.manager [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] No waiting events found dispatching network-vif-plugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 778.752998] env[61768]: WARNING nova.compute.manager [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received unexpected event network-vif-plugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 for instance with vm_state building and task_state spawning. [ 778.753314] env[61768]: DEBUG nova.compute.manager [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 778.753523] env[61768]: DEBUG nova.compute.manager [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing instance network info cache due to event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 778.753746] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.753896] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.754105] env[61768]: DEBUG nova.network.neutron [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 778.755476] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5934f0b0-7d84-4af7-a6d9-7f8eb370bfcf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.787658] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228805, 'name': Rename_Task, 'duration_secs': 0.15885} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.793390] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 778.793390] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228803, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.053128} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.793390] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 778.793390] env[61768]: value = "task-1228807" [ 778.793390] env[61768]: _type = "Task" [ 778.793390] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.793390] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-89d4672c-4fbb-43df-8790-54b485353d01 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.794681] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 778.794811] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 778.795856] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aea43be3-6a27-46b6-85af-dec82ea9f4b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.806630] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228807, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.808677] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 778.808677] env[61768]: value = "task-1228809" [ 778.808677] env[61768]: _type = "Task" [ 778.808677] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.809054] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 778.809054] env[61768]: value = "task-1228808" [ 778.809054] env[61768]: _type = "Task" [ 778.809054] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.823786] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228809, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.827838] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228808, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.092471] env[61768]: DEBUG nova.network.neutron [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.144900] env[61768]: DEBUG nova.scheduler.client.report [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.236416] env[61768]: DEBUG oslo_vmware.api [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228806, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.252588] env[61768]: DEBUG nova.network.neutron [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updated VIF entry in instance network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 779.252959] env[61768]: DEBUG nova.network.neutron [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.306136] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228807, 'name': ReconfigVM_Task, 'duration_secs': 0.458849} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.307478] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 77319462-e447-405b-9269-82581effe005/77319462-e447-405b-9269-82581effe005.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 779.308918] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfb26417-60a9-476d-b435-4798324aa0a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.330727] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228809, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.232401} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.331236] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228808, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.333388] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 779.335109] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 779.335109] env[61768]: value = "task-1228810" [ 779.335109] env[61768]: _type = "Task" [ 779.335109] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.335109] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9555add-520a-4bc5-98da-a8adb3222ec9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.368529] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 779.371238] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd20ca3f-e7a9-42c4-82a8-ee799c16e8d5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.391157] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228810, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.398981] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 779.398981] env[61768]: value = "task-1228811" [ 779.398981] env[61768]: _type = "Task" [ 779.398981] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.409502] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228811, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.442894] env[61768]: DEBUG nova.network.neutron [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Updating instance_info_cache with network_info: [{"id": "49a82793-2751-49db-b262-593b386ea68d", "address": "fa:16:3e:67:7c:ab", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49a82793-27", "ovs_interfaceid": "49a82793-2751-49db-b262-593b386ea68d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.572824] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "76c282d1-ddfe-46dc-aa7b-225708443379" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.573212] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "76c282d1-ddfe-46dc-aa7b-225708443379" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.573500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "76c282d1-ddfe-46dc-aa7b-225708443379-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.573973] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "76c282d1-ddfe-46dc-aa7b-225708443379-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.574077] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "76c282d1-ddfe-46dc-aa7b-225708443379-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.576810] env[61768]: INFO nova.compute.manager [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Terminating instance [ 779.578880] env[61768]: DEBUG nova.compute.manager [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 779.579109] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 779.580214] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecbe9ee2-9824-4cdd-ae6e-a04695e6cdbf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.589316] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 779.589635] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e6bc7a8-2bf2-46f5-9c0e-f620800c5271 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.600866] env[61768]: DEBUG oslo_vmware.api [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 779.600866] env[61768]: value = "task-1228812" [ 779.600866] env[61768]: _type = "Task" [ 779.600866] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.609331] env[61768]: DEBUG oslo_vmware.api [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.635747] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received event network-changed-af1a7711-f21e-4462-93ef-a052655ce43c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 779.636136] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 77319462-e447-405b-9269-82581effe005] Refreshing instance network info cache due to event network-changed-af1a7711-f21e-4462-93ef-a052655ce43c. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 779.637046] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquiring lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.637046] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquired lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.637046] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 77319462-e447-405b-9269-82581effe005] Refreshing network info cache for port af1a7711-f21e-4462-93ef-a052655ce43c {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 779.653811] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.654535] env[61768]: DEBUG nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.659291] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.095s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.659636] env[61768]: DEBUG nova.objects.instance [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lazy-loading 'resources' on Instance uuid efbde802-8cb0-4563-a776-3722a3889afe {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.732689] env[61768]: DEBUG oslo_vmware.api [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228806, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.756399] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb630476-cfe4-452f-b60c-ca8b0dadeb9b req-5c194f93-cfb6-403e-aac0-471d20b474b9 service nova] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.823336] env[61768]: DEBUG oslo_vmware.api [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228808, 'name': PowerOnVM_Task, 'duration_secs': 0.597589} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.824517] env[61768]: DEBUG nova.network.neutron [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Successfully updated port: 3bc689b6-7278-4741-8053-b98ed2db0265 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 779.826031] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 779.826433] env[61768]: INFO nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Took 9.06 seconds to spawn the instance on the hypervisor. [ 779.826749] env[61768]: DEBUG nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 779.828273] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0febf7c9-75ec-4cca-981c-8d6ec805cdc3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.853490] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228810, 'name': Rename_Task, 'duration_secs': 0.334598} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.854082] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 779.854175] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-763d02f4-a0ff-419f-addc-4d7d397a95e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.863658] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 779.863658] env[61768]: value = "task-1228813" [ 779.863658] env[61768]: _type = "Task" [ 779.863658] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.875904] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228813, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.913689] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228811, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.946520] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.946905] env[61768]: DEBUG nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Instance network_info: |[{"id": "49a82793-2751-49db-b262-593b386ea68d", "address": "fa:16:3e:67:7c:ab", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49a82793-27", "ovs_interfaceid": "49a82793-2751-49db-b262-593b386ea68d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 779.947761] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:67:7c:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '49a82793-2751-49db-b262-593b386ea68d', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.958481] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating folder: Project (c7857b5217cc499ca8c9c3b8a5471db7). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 779.959193] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-477ca57b-eee8-462a-9104-ce2f521fc3aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.974902] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created folder: Project (c7857b5217cc499ca8c9c3b8a5471db7) in parent group-v265360. [ 779.975154] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating folder: Instances. Parent ref: group-v265436. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 779.976083] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a489e82f-1507-47e5-817c-6eea136da79e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.988819] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created folder: Instances in parent group-v265436. [ 779.991146] env[61768]: DEBUG oslo.service.loopingcall [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.991146] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 779.991146] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3fa5eb03-1514-4c8c-8f0c-99a0a78318e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.018953] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.018953] env[61768]: value = "task-1228816" [ 780.018953] env[61768]: _type = "Task" [ 780.018953] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.029915] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228816, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.111905] env[61768]: DEBUG oslo_vmware.api [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228812, 'name': PowerOffVM_Task, 'duration_secs': 0.386139} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.111905] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 780.111905] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 780.111905] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c93b4a6b-edb1-42eb-a974-3a62feb680fc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.165596] env[61768]: DEBUG nova.compute.utils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.167020] env[61768]: DEBUG nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.167182] env[61768]: DEBUG nova.network.neutron [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.239502] env[61768]: DEBUG oslo_vmware.api [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228806, 'name': ReconfigVM_Task, 'duration_secs': 1.026455} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.241250] env[61768]: DEBUG nova.policy [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7f7389d526c4ce9bc8ecf062b9921e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '180da0bcd5954eb2bc89a20cf8f42bb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.243602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.243830] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfigured VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 780.304158] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 780.304500] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 780.305225] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleting the datastore file [datastore1] 76c282d1-ddfe-46dc-aa7b-225708443379 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.305225] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96541a59-87f7-4f89-bbbd-b76e380036f5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.315825] env[61768]: DEBUG oslo_vmware.api [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 780.315825] env[61768]: value = "task-1228818" [ 780.315825] env[61768]: _type = "Task" [ 780.315825] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.326056] env[61768]: DEBUG oslo_vmware.api [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228818, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.328867] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "refresh_cache-fbb8c9db-e823-444a-84de-b231cf9b8a8a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.328992] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquired lock "refresh_cache-fbb8c9db-e823-444a-84de-b231cf9b8a8a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.329334] env[61768]: DEBUG nova.network.neutron [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.354761] env[61768]: INFO nova.compute.manager [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Took 33.76 seconds to build instance. [ 780.381705] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228813, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.411350] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228811, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.532254] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228816, 'name': CreateVM_Task, 'duration_secs': 0.460717} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.532497] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 780.533810] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.533810] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.534894] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 780.540531] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81693248-6c0e-4035-8867-05f6a05b2d55 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.551296] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 780.551296] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f1cf1e-b9b6-4f41-3dc9-b3fa79ee9e5b" [ 780.551296] env[61768]: _type = "Task" [ 780.551296] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.566609] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f1cf1e-b9b6-4f41-3dc9-b3fa79ee9e5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.568898] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 77319462-e447-405b-9269-82581effe005] Updated VIF entry in instance network info cache for port af1a7711-f21e-4462-93ef-a052655ce43c. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 780.569062] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 77319462-e447-405b-9269-82581effe005] Updating instance_info_cache with network_info: [{"id": "5c1feaf9-e399-4171-90ae-e50db59e5274", "address": "fa:16:3e:90:0b:b0", "network": {"id": "936f4d95-f203-43f8-ba39-ef22c1c8ea4e", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1357034332", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.83", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "60badc2d-69d2-467d-a92e-98511f5cb0b2", "external-id": "cl2-zone-408", "segmentation_id": 408, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5c1feaf9-e3", "ovs_interfaceid": "5c1feaf9-e399-4171-90ae-e50db59e5274", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "af1a7711-f21e-4462-93ef-a052655ce43c", "address": "fa:16:3e:e3:57:9b", "network": {"id": "d3307f81-acee-4479-a53a-a0c40369ec9f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1851008898", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "ecf64533220b4a5b9f05c0c864e408a0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c4349e30-c086-4c24-9e0e-83996d808a1b", "external-id": "nsx-vlan-transportzone-266", "segmentation_id": 266, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaf1a7711-f2", "ovs_interfaceid": "af1a7711-f21e-4462-93ef-a052655ce43c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.570995] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc13671-e23d-4700-b27f-16169e59ae63 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.581864] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf27d2f-1dc0-4cac-a0ba-bcd5f9a91077 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.619076] env[61768]: DEBUG nova.network.neutron [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Successfully created port: 61d36258-9772-43d5-aa74-55c86e1596a7 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.621892] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63655873-e833-4825-80f9-fb85bae97a39 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.633100] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f931df5-1d99-4ef5-b11a-53cee0c636a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.651892] env[61768]: DEBUG nova.compute.provider_tree [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.671463] env[61768]: DEBUG nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.753347] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8408405-8f9c-4453-82e4-376d5e250fd1 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-cb8850cf-e632-419a-86f0-4c532e443213" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.682s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.829734] env[61768]: DEBUG oslo_vmware.api [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228818, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222289} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.830108] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 780.830374] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 780.830631] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 780.830887] env[61768]: INFO nova.compute.manager [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Took 1.25 seconds to destroy the instance on the hypervisor. [ 780.832161] env[61768]: DEBUG oslo.service.loopingcall [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 780.834957] env[61768]: DEBUG nova.compute.manager [-] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.835126] env[61768]: DEBUG nova.network.neutron [-] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 780.856978] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1d8e4c83-d169-43f6-aea2-eba359e9c14f tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.699s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.870303] env[61768]: DEBUG nova.network.neutron [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 780.878346] env[61768]: DEBUG oslo_vmware.api [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228813, 'name': PowerOnVM_Task, 'duration_secs': 0.670969} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.879511] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 780.879511] env[61768]: INFO nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Took 12.59 seconds to spawn the instance on the hypervisor. [ 780.879647] env[61768]: DEBUG nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 780.880507] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2867952e-3f08-46f1-b5a9-b45e80a56d25 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.916622] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228811, 'name': ReconfigVM_Task, 'duration_secs': 1.243876} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.916931] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 780.918325] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fa9fad3f-8274-4f9c-921f-7efb03a4f64f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.926940] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 780.926940] env[61768]: value = "task-1228819" [ 780.926940] env[61768]: _type = "Task" [ 780.926940] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.940804] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228819, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.061701] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f1cf1e-b9b6-4f41-3dc9-b3fa79ee9e5b, 'name': SearchDatastore_Task, 'duration_secs': 0.01968} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.062227] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.062628] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.063170] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.063466] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.063775] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.064199] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-41fa7fb0-3180-4115-9555-4fa1e7ae2e62 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.076242] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Releasing lock "refresh_cache-77319462-e447-405b-9269-82581effe005" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.076522] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-vif-plugged-cb8850cf-e632-419a-86f0-4c532e443213 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.076811] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.077065] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.077246] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.077438] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] No waiting events found dispatching network-vif-plugged-cb8850cf-e632-419a-86f0-4c532e443213 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 781.077698] env[61768]: WARNING nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received unexpected event network-vif-plugged-cb8850cf-e632-419a-86f0-4c532e443213 for instance with vm_state active and task_state None. [ 781.077928] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.078127] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing instance network info cache due to event network-changed-61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 781.078362] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquiring lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.078533] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquired lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.078726] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Refreshing network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 781.081767] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.081767] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 781.082596] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e6b095a-9b1b-497a-99e5-686f2476e36a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.090259] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 781.090259] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a75520-9f6d-9fc2-c79d-892fcb0b6448" [ 781.090259] env[61768]: _type = "Task" [ 781.090259] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.102306] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a75520-9f6d-9fc2-c79d-892fcb0b6448, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.107829] env[61768]: DEBUG nova.network.neutron [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Updating instance_info_cache with network_info: [{"id": "3bc689b6-7278-4741-8053-b98ed2db0265", "address": "fa:16:3e:47:ae:aa", "network": {"id": "c7ac5eb3-a893-4e46-9681-340ef208d79b", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1207890730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f595b30e898b426a9675202e54338a4b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bc689b6-72", "ovs_interfaceid": "3bc689b6-7278-4741-8053-b98ed2db0265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.155507] env[61768]: DEBUG nova.scheduler.client.report [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.362222] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.400992] env[61768]: INFO nova.compute.manager [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Took 35.36 seconds to build instance. [ 781.441024] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228819, 'name': Rename_Task, 'duration_secs': 0.235579} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.441522] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 781.441896] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d16d030b-8328-46ec-9c71-cdc7237a7f7e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.449508] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 781.449508] env[61768]: value = "task-1228820" [ 781.449508] env[61768]: _type = "Task" [ 781.449508] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.458706] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228820, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.516102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "d4a765bd-a56d-44b7-8db3-c081832d58c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.516102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.544310] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "3aa722e5-0818-40ae-8220-223d920a7386" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.544616] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "3aa722e5-0818-40ae-8220-223d920a7386" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.604872] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a75520-9f6d-9fc2-c79d-892fcb0b6448, 'name': SearchDatastore_Task, 'duration_secs': 0.014234} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.605692] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8756ec7d-b81d-4dfb-a4a1-230a53b48ea9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.611623] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Releasing lock "refresh_cache-fbb8c9db-e823-444a-84de-b231cf9b8a8a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.611924] env[61768]: DEBUG nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Instance network_info: |[{"id": "3bc689b6-7278-4741-8053-b98ed2db0265", "address": "fa:16:3e:47:ae:aa", "network": {"id": "c7ac5eb3-a893-4e46-9681-340ef208d79b", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1207890730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f595b30e898b426a9675202e54338a4b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bc689b6-72", "ovs_interfaceid": "3bc689b6-7278-4741-8053-b98ed2db0265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 781.612308] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 781.612308] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a5df5a-63e4-ff02-ca87-97f7f275f609" [ 781.612308] env[61768]: _type = "Task" [ 781.612308] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.614869] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:ae:aa', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '664c466b-9417-49d7-83cc-364d964c403a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3bc689b6-7278-4741-8053-b98ed2db0265', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 781.623363] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Creating folder: Project (f595b30e898b426a9675202e54338a4b). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.623363] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e96a847e-42e7-429c-acde-cd5f979a329b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.633762] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a5df5a-63e4-ff02-ca87-97f7f275f609, 'name': SearchDatastore_Task, 'duration_secs': 0.010717} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.634058] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.634334] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 2c85de5d-0111-40a6-a8bc-69c7eba0393d/2c85de5d-0111-40a6-a8bc-69c7eba0393d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 781.634607] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-700fb000-ee65-48b3-8400-48c5d90dce95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.637838] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Created folder: Project (f595b30e898b426a9675202e54338a4b) in parent group-v265360. [ 781.638039] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Creating folder: Instances. Parent ref: group-v265439. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 781.638589] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0d4be6c-ac1a-459d-aeda-b9996717cb41 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.643733] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 781.643733] env[61768]: value = "task-1228822" [ 781.643733] env[61768]: _type = "Task" [ 781.643733] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.648877] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Created folder: Instances in parent group-v265439. [ 781.648992] env[61768]: DEBUG oslo.service.loopingcall [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 781.649163] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 781.652574] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52cce729-ec5d-4734-bb21-f83cebec1835 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.667488] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.668242] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.009s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.670348] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.032s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.671974] env[61768]: INFO nova.compute.claims [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.681602] env[61768]: DEBUG nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.683700] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 781.683700] env[61768]: value = "task-1228824" [ 781.683700] env[61768]: _type = "Task" [ 781.683700] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.687311] env[61768]: DEBUG nova.network.neutron [-] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.697710] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228824, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.698967] env[61768]: INFO nova.scheduler.client.report [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted allocations for instance efbde802-8cb0-4563-a776-3722a3889afe [ 781.715990] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.716395] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.716606] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.716839] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.717035] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.717246] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.717522] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.717799] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.717947] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.718741] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.718741] env[61768]: DEBUG nova.virt.hardware [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.720093] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29c99bb-43c4-47d2-be2e-112132d2829e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.732648] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e0d173-7fe3-44f0-8312-b73d38a48388 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.859441] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updated VIF entry in instance network info cache for port 61e8c28f-c389-4d6d-af8e-dabbc4c83363. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 781.859441] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [{"id": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "address": "fa:16:3e:98:b7:f8", "network": {"id": "3e4a66df-17f2-479e-bb37-f8659f0180d6", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-1225762599-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "ec5c061ca09040e8a2b3c3a6874daa40", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbf3349e-d05e-4d44-a011-c4b6e41af988", "external-id": "nsx-vlan-transportzone-196", "segmentation_id": 196, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61e8c28f-c3", "ovs_interfaceid": "61e8c28f-c389-4d6d-af8e-dabbc4c83363", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.889455] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.907032] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2cd10a89-f187-4624-b49a-6cf93dedee32 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "77319462-e447-405b-9269-82581effe005" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.475s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.967097] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228820, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.973819] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "23a4450d-162a-4b2b-a009-7023851315a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.973819] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "23a4450d-162a-4b2b-a009-7023851315a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.973819] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "23a4450d-162a-4b2b-a009-7023851315a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.973819] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "23a4450d-162a-4b2b-a009-7023851315a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.973819] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "23a4450d-162a-4b2b-a009-7023851315a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.975042] env[61768]: INFO nova.compute.manager [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Terminating instance [ 781.977628] env[61768]: DEBUG nova.compute.manager [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 781.977849] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 781.978763] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbbb26a1-58ae-4d5c-afd2-c548110161e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.988462] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 781.988833] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80c86742-989a-464e-a3d9-17ccb92136c9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.999062] env[61768]: DEBUG oslo_vmware.api [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 781.999062] env[61768]: value = "task-1228825" [ 781.999062] env[61768]: _type = "Task" [ 781.999062] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.014801] env[61768]: DEBUG oslo_vmware.api [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.156985] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228822, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.185462] env[61768]: DEBUG nova.compute.manager [req-6c06bb9c-cc27-4d8e-8376-eb22d7207cc0 req-45519602-844f-4a27-93b6-077197c630bf service nova] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Received event network-vif-deleted-b4dfbd11-5359-4bf8-bd2f-f2704d5ead57 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.193153] env[61768]: INFO nova.compute.manager [-] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Took 1.36 seconds to deallocate network for instance. [ 782.208557] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228824, 'name': CreateVM_Task, 'duration_secs': 0.422967} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.210134] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 782.211409] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75aebe7a-64c1-4cf8-9aea-194a66cbb2f1 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "efbde802-8cb0-4563-a776-3722a3889afe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.526s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.212869] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.213061] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.213592] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 782.214684] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a8fb86b-d296-4129-8b11-0f5c8be90d62 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.224263] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 782.224263] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5212dedf-8816-402a-96b6-3787f5be72f6" [ 782.224263] env[61768]: _type = "Task" [ 782.224263] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.237777] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5212dedf-8816-402a-96b6-3787f5be72f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.333105] env[61768]: DEBUG nova.network.neutron [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Successfully updated port: 61d36258-9772-43d5-aa74-55c86e1596a7 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.362730] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Releasing lock "refresh_cache-23a4450d-162a-4b2b-a009-7023851315a1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.363181] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-changed-cb8850cf-e632-419a-86f0-4c532e443213 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.363408] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing instance network info cache due to event network-changed-cb8850cf-e632-419a-86f0-4c532e443213. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.363760] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.363849] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.364037] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Refreshing network info cache for port cb8850cf-e632-419a-86f0-4c532e443213 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.402904] env[61768]: DEBUG nova.compute.manager [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Received event network-vif-plugged-3bc689b6-7278-4741-8053-b98ed2db0265 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.403146] env[61768]: DEBUG oslo_concurrency.lockutils [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] Acquiring lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.403454] env[61768]: DEBUG oslo_concurrency.lockutils [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.403673] env[61768]: DEBUG oslo_concurrency.lockutils [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.403887] env[61768]: DEBUG nova.compute.manager [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] No waiting events found dispatching network-vif-plugged-3bc689b6-7278-4741-8053-b98ed2db0265 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 782.404243] env[61768]: WARNING nova.compute.manager [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Received unexpected event network-vif-plugged-3bc689b6-7278-4741-8053-b98ed2db0265 for instance with vm_state building and task_state spawning. [ 782.404556] env[61768]: DEBUG nova.compute.manager [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Received event network-changed-3bc689b6-7278-4741-8053-b98ed2db0265 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 782.404556] env[61768]: DEBUG nova.compute.manager [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Refreshing instance network info cache due to event network-changed-3bc689b6-7278-4741-8053-b98ed2db0265. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 782.404704] env[61768]: DEBUG oslo_concurrency.lockutils [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] Acquiring lock "refresh_cache-fbb8c9db-e823-444a-84de-b231cf9b8a8a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.404836] env[61768]: DEBUG oslo_concurrency.lockutils [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] Acquired lock "refresh_cache-fbb8c9db-e823-444a-84de-b231cf9b8a8a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.405052] env[61768]: DEBUG nova.network.neutron [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Refreshing network info cache for port 3bc689b6-7278-4741-8053-b98ed2db0265 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.408228] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 782.462399] env[61768]: DEBUG oslo_vmware.api [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228820, 'name': PowerOnVM_Task, 'duration_secs': 0.814688} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.462715] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 782.462927] env[61768]: INFO nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Took 9.22 seconds to spawn the instance on the hypervisor. [ 782.463126] env[61768]: DEBUG nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 782.464130] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b390478-7680-44fd-942c-ea0ae5d82c5e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.510250] env[61768]: DEBUG oslo_vmware.api [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228825, 'name': PowerOffVM_Task, 'duration_secs': 0.340984} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.510480] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 782.510682] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 782.513736] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-80390ee7-ec01-46fd-b9bd-33cbac3f0f5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.598167] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 782.598167] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 782.598167] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Deleting the datastore file [datastore2] 23a4450d-162a-4b2b-a009-7023851315a1 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.598167] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4d15e01e-d488-44fd-9ea2-acbdec8238b8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.605998] env[61768]: DEBUG oslo_vmware.api [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for the task: (returnval){ [ 782.605998] env[61768]: value = "task-1228827" [ 782.605998] env[61768]: _type = "Task" [ 782.605998] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.617082] env[61768]: DEBUG oslo_vmware.api [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.656187] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572042} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.656505] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 2c85de5d-0111-40a6-a8bc-69c7eba0393d/2c85de5d-0111-40a6-a8bc-69c7eba0393d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 782.656775] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 782.657089] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-273bc78c-9fe9-4a1d-bdc1-7c1da09d5d57 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.670135] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1339579-c392-4725-9b33-c0e6bf0be676 tempest-ServersListShow296Test-1960977358 tempest-ServersListShow296Test-1960977358-project-member] Acquiring lock "54d8c044-3cc6-4e0c-a809-cc8351b80b27" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.670413] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1339579-c392-4725-9b33-c0e6bf0be676 tempest-ServersListShow296Test-1960977358 tempest-ServersListShow296Test-1960977358-project-member] Lock "54d8c044-3cc6-4e0c-a809-cc8351b80b27" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.670731] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 782.670731] env[61768]: value = "task-1228828" [ 782.670731] env[61768]: _type = "Task" [ 782.670731] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.681675] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228828, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.706890] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.738690] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5212dedf-8816-402a-96b6-3787f5be72f6, 'name': SearchDatastore_Task, 'duration_secs': 0.014191} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.738999] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.739262] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 782.739508] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.739692] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.739844] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 782.740533] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec2aeb6c-b74d-48bf-b6fd-6e71767aaa7b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.750433] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 782.750433] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 782.751191] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f90f3ab0-67b2-4198-94fa-96240f69f9f7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.757107] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 782.757107] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5266b976-d222-29af-1270-9c292f603aa7" [ 782.757107] env[61768]: _type = "Task" [ 782.757107] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.768472] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5266b976-d222-29af-1270-9c292f603aa7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.836672] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "refresh_cache-3d96e47d-fc09-439e-bf0e-e60fc87ae958" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.836831] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "refresh_cache-3d96e47d-fc09-439e-bf0e-e60fc87ae958" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.836987] env[61768]: DEBUG nova.network.neutron [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.932269] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.987317] env[61768]: INFO nova.compute.manager [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Took 32.75 seconds to build instance. [ 782.997194] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6f7b1f3-89a6-491d-b8c3-040575c679d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.005956] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b608f0e6-8127-491a-9d6c-5942619e9807 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.039542] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59aa5897-3d6c-4d9b-983e-d7a295691eb1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.049034] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6df72d1-156e-4c16-9723-721de1332d50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.070146] env[61768]: DEBUG nova.compute.provider_tree [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.121203] env[61768]: DEBUG oslo_vmware.api [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.185030] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228828, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076316} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.185030] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 783.185030] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de062cb-690a-4a3e-a7e1-600309f52c84 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.209438] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 2c85de5d-0111-40a6-a8bc-69c7eba0393d/2c85de5d-0111-40a6-a8bc-69c7eba0393d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 783.212155] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bf90fce0-30ee-4453-af2f-a12327286256 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.239080] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 783.239080] env[61768]: value = "task-1228829" [ 783.239080] env[61768]: _type = "Task" [ 783.239080] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.250205] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228829, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.277230] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5266b976-d222-29af-1270-9c292f603aa7, 'name': SearchDatastore_Task, 'duration_secs': 0.015478} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.278073] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-441025a3-de5e-43de-a7a6-5f0ee2808db7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.284640] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 783.284640] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527795d2-89b5-9a41-9de4-6a5db4e4ade1" [ 783.284640] env[61768]: _type = "Task" [ 783.284640] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.292746] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527795d2-89b5-9a41-9de4-6a5db4e4ade1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.380504] env[61768]: DEBUG nova.network.neutron [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.490333] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f3bb2aa2-0266-4f4b-88bb-3f2679fe8539 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 65.052s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.515464] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updated VIF entry in instance network info cache for port cb8850cf-e632-419a-86f0-4c532e443213. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.515792] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "address": "fa:16:3e:7d:36:19", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa0589e32-42", "ovs_interfaceid": "a0589e32-427b-4fea-a7f6-6de811fcebdb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cb8850cf-e632-419a-86f0-4c532e443213", "address": "fa:16:3e:c8:6d:50", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb8850cf-e6", "ovs_interfaceid": "cb8850cf-e632-419a-86f0-4c532e443213", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.518150] env[61768]: DEBUG nova.network.neutron [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Updated VIF entry in instance network info cache for port 3bc689b6-7278-4741-8053-b98ed2db0265. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.518474] env[61768]: DEBUG nova.network.neutron [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Updating instance_info_cache with network_info: [{"id": "3bc689b6-7278-4741-8053-b98ed2db0265", "address": "fa:16:3e:47:ae:aa", "network": {"id": "c7ac5eb3-a893-4e46-9681-340ef208d79b", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1207890730-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f595b30e898b426a9675202e54338a4b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "664c466b-9417-49d7-83cc-364d964c403a", "external-id": "nsx-vlan-transportzone-103", "segmentation_id": 103, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3bc689b6-72", "ovs_interfaceid": "3bc689b6-7278-4741-8053-b98ed2db0265", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.573514] env[61768]: DEBUG nova.scheduler.client.report [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.597309] env[61768]: DEBUG nova.network.neutron [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Updating instance_info_cache with network_info: [{"id": "61d36258-9772-43d5-aa74-55c86e1596a7", "address": "fa:16:3e:4c:2d:3b", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61d36258-97", "ovs_interfaceid": "61d36258-9772-43d5-aa74-55c86e1596a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.620501] env[61768]: DEBUG oslo_vmware.api [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Task: {'id': task-1228827, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.663614} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.620637] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 783.620763] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 783.620946] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 783.621178] env[61768]: INFO nova.compute.manager [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Took 1.64 seconds to destroy the instance on the hypervisor. [ 783.621458] env[61768]: DEBUG oslo.service.loopingcall [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 783.621684] env[61768]: DEBUG nova.compute.manager [-] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 783.621795] env[61768]: DEBUG nova.network.neutron [-] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.750172] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228829, 'name': ReconfigVM_Task, 'duration_secs': 0.351423} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.750312] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 2c85de5d-0111-40a6-a8bc-69c7eba0393d/2c85de5d-0111-40a6-a8bc-69c7eba0393d.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 783.750946] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c05450a9-1f05-4459-932a-484bde8c6b8b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.759235] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 783.759235] env[61768]: value = "task-1228830" [ 783.759235] env[61768]: _type = "Task" [ 783.759235] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.769953] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228830, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.798024] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527795d2-89b5-9a41-9de4-6a5db4e4ade1, 'name': SearchDatastore_Task, 'duration_secs': 0.017806} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.798024] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.798024] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] fbb8c9db-e823-444a-84de-b231cf9b8a8a/fbb8c9db-e823-444a-84de-b231cf9b8a8a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 783.798024] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9dd2a05-1752-4c8e-a2d0-cf34d138e38f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.804439] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 783.804439] env[61768]: value = "task-1228831" [ 783.804439] env[61768]: _type = "Task" [ 783.804439] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.814165] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.996018] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 784.020559] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.020965] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Received event network-vif-plugged-49a82793-2751-49db-b262-593b386ea68d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.021248] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquiring lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.021490] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.021692] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.021932] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] No waiting events found dispatching network-vif-plugged-49a82793-2751-49db-b262-593b386ea68d {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.022301] env[61768]: WARNING nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Received unexpected event network-vif-plugged-49a82793-2751-49db-b262-593b386ea68d for instance with vm_state building and task_state spawning. [ 784.022389] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Received event network-changed-49a82793-2751-49db-b262-593b386ea68d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.022968] env[61768]: DEBUG nova.compute.manager [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Refreshing instance network info cache due to event network-changed-49a82793-2751-49db-b262-593b386ea68d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 784.022968] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquiring lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.023117] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Acquired lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.023317] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Refreshing network info cache for port 49a82793-2751-49db-b262-593b386ea68d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.025512] env[61768]: DEBUG oslo_concurrency.lockutils [req-34c2e0ec-bf12-456a-99d9-e86ecaf532fb req-fc4db09d-a9ba-4546-a60b-d3e74528be2a service nova] Releasing lock "refresh_cache-fbb8c9db-e823-444a-84de-b231cf9b8a8a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.082046] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.082046] env[61768]: DEBUG nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.085476] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.706s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.086061] env[61768]: DEBUG nova.objects.instance [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lazy-loading 'resources' on Instance uuid d70040f1-e425-4aeb-b82d-ce808a02a645 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 784.100380] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "refresh_cache-3d96e47d-fc09-439e-bf0e-e60fc87ae958" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.101053] env[61768]: DEBUG nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Instance network_info: |[{"id": "61d36258-9772-43d5-aa74-55c86e1596a7", "address": "fa:16:3e:4c:2d:3b", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61d36258-97", "ovs_interfaceid": "61d36258-9772-43d5-aa74-55c86e1596a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 784.101622] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:2d:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61d36258-9772-43d5-aa74-55c86e1596a7', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 784.110916] env[61768]: DEBUG oslo.service.loopingcall [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.111488] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 784.111698] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fd4a50d-e84c-4d53-87a4-d44cead929b0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.134907] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 784.134907] env[61768]: value = "task-1228832" [ 784.134907] env[61768]: _type = "Task" [ 784.134907] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.143859] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228832, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.225418] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "77319462-e447-405b-9269-82581effe005" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.226092] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "77319462-e447-405b-9269-82581effe005" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.226455] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "77319462-e447-405b-9269-82581effe005-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.226844] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "77319462-e447-405b-9269-82581effe005-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.227223] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "77319462-e447-405b-9269-82581effe005-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.231959] env[61768]: INFO nova.compute.manager [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Terminating instance [ 784.235369] env[61768]: DEBUG nova.compute.manager [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.236460] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 784.239586] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48cbd9f-1d01-440e-a4a2-31a8fc2d92cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.256494] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 784.256886] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0923f286-e37d-4b9b-a64e-640c567f7d29 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.281142] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 784.281142] env[61768]: value = "task-1228833" [ 784.281142] env[61768]: _type = "Task" [ 784.281142] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.281142] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228830, 'name': Rename_Task, 'duration_secs': 0.144619} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.281142] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 784.286070] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7667506-28e8-4454-b636-1e669c77c448 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.296905] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228833, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.298938] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 784.298938] env[61768]: value = "task-1228834" [ 784.298938] env[61768]: _type = "Task" [ 784.298938] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.312844] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228834, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.322109] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228831, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.403387] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-a0589e32-427b-4fea-a7f6-6de811fcebdb" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.403387] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-a0589e32-427b-4fea-a7f6-6de811fcebdb" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.427016] env[61768]: DEBUG nova.compute.manager [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Received event network-vif-plugged-61d36258-9772-43d5-aa74-55c86e1596a7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.427672] env[61768]: DEBUG oslo_concurrency.lockutils [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] Acquiring lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.428317] env[61768]: DEBUG oslo_concurrency.lockutils [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.428718] env[61768]: DEBUG oslo_concurrency.lockutils [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.429204] env[61768]: DEBUG nova.compute.manager [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] No waiting events found dispatching network-vif-plugged-61d36258-9772-43d5-aa74-55c86e1596a7 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 784.429597] env[61768]: WARNING nova.compute.manager [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Received unexpected event network-vif-plugged-61d36258-9772-43d5-aa74-55c86e1596a7 for instance with vm_state building and task_state spawning. [ 784.430165] env[61768]: DEBUG nova.compute.manager [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Received event network-changed-61d36258-9772-43d5-aa74-55c86e1596a7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.432023] env[61768]: DEBUG nova.compute.manager [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Refreshing instance network info cache due to event network-changed-61d36258-9772-43d5-aa74-55c86e1596a7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 784.432023] env[61768]: DEBUG oslo_concurrency.lockutils [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] Acquiring lock "refresh_cache-3d96e47d-fc09-439e-bf0e-e60fc87ae958" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.432023] env[61768]: DEBUG oslo_concurrency.lockutils [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] Acquired lock "refresh_cache-3d96e47d-fc09-439e-bf0e-e60fc87ae958" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.432023] env[61768]: DEBUG nova.network.neutron [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Refreshing network info cache for port 61d36258-9772-43d5-aa74-55c86e1596a7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.440808] env[61768]: DEBUG nova.network.neutron [-] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.524324] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.592826] env[61768]: DEBUG nova.compute.utils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.597249] env[61768]: DEBUG nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 784.597788] env[61768]: DEBUG nova.network.neutron [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 784.652161] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228832, 'name': CreateVM_Task, 'duration_secs': 0.471784} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.652364] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 784.653154] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.653341] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.654301] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 784.654637] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3620a2a-2508-4f46-b14a-fe350062bf0f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.664208] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 784.664208] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523392a4-88df-2aa1-c64c-76f873630439" [ 784.664208] env[61768]: _type = "Task" [ 784.664208] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.674528] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523392a4-88df-2aa1-c64c-76f873630439, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.698026] env[61768]: DEBUG nova.policy [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9d7f72b4c22452d89004ca70d165e67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f4804c086b449608ef34ea22c018907', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.735417] env[61768]: DEBUG nova.compute.manager [req-b121a3d0-d21f-48e9-aacf-2a0ee50682a5 req-2e6e6284-a149-433a-aba1-f9751db3e5f4 service nova] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Received event network-vif-deleted-61e8c28f-c389-4d6d-af8e-dabbc4c83363 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 784.797013] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228833, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.815934] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228834, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.826330] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.733276} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.826709] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] fbb8c9db-e823-444a-84de-b231cf9b8a8a/fbb8c9db-e823-444a-84de-b231cf9b8a8a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 784.826982] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.827316] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c3b6ec92-f788-4cdd-b1ad-b0be046e856d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.839444] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 784.839444] env[61768]: value = "task-1228835" [ 784.839444] env[61768]: _type = "Task" [ 784.839444] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.850400] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228835, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.860885] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.861763] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.861763] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.861763] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.861997] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.864688] env[61768]: INFO nova.compute.manager [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Terminating instance [ 784.867177] env[61768]: DEBUG nova.compute.manager [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.867292] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 784.868216] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574b36d3-d3a7-4b42-821a-3a62609cc5e3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.882747] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 784.886569] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-035c8110-4fee-46cb-b62c-c05d3460baa3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.899839] env[61768]: DEBUG oslo_vmware.api [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 784.899839] env[61768]: value = "task-1228836" [ 784.899839] env[61768]: _type = "Task" [ 784.899839] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.909633] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.909805] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.911724] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e7909d-0b8b-4baf-89a3-be4304605d40 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.925128] env[61768]: DEBUG oslo_vmware.api [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.945386] env[61768]: INFO nova.compute.manager [-] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Took 1.32 seconds to deallocate network for instance. [ 784.950180] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff69920f-f050-4523-876f-dab5b9905ac7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.955164] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Updated VIF entry in instance network info cache for port 49a82793-2751-49db-b262-593b386ea68d. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 784.955628] env[61768]: DEBUG nova.network.neutron [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Updating instance_info_cache with network_info: [{"id": "49a82793-2751-49db-b262-593b386ea68d", "address": "fa:16:3e:67:7c:ab", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49a82793-27", "ovs_interfaceid": "49a82793-2751-49db-b262-593b386ea68d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.989677] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfiguring VM to detach interface {{(pid=61768) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 784.994101] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a34d7c32-6283-4ca5-9b9c-efe586d88849 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.024025] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 785.024025] env[61768]: value = "task-1228837" [ 785.024025] env[61768]: _type = "Task" [ 785.024025] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.035583] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81156795-8ee7-45b4-a0a0-264170206d42 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.042742] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.050747] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb5bacf-e1df-42cf-92ef-7b2e9abe0ebb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.093063] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dad139-6168-4099-a07d-16ff98a5b98d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.102441] env[61768]: DEBUG nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.110025] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a635b3c7-6e8c-4374-b00c-886f3015e08a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.126700] env[61768]: DEBUG nova.compute.provider_tree [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.179244] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523392a4-88df-2aa1-c64c-76f873630439, 'name': SearchDatastore_Task, 'duration_secs': 0.015042} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.179244] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.179244] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 785.179244] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.179244] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.179244] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 785.179244] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a042233b-309d-4a76-b924-1ddda916d705 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.191466] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 785.191664] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 785.192442] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4383943-14e9-4299-a41d-aaf3145dd005 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.204062] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 785.204062] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a2dec-08d9-66a5-4e0e-6299d3d44f80" [ 785.204062] env[61768]: _type = "Task" [ 785.204062] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.211189] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a2dec-08d9-66a5-4e0e-6299d3d44f80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.294419] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228833, 'name': PowerOffVM_Task, 'duration_secs': 0.610565} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.294879] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 785.294879] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 785.295202] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4d4e7ef-5c38-4649-9ecd-3cf99d425fad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.312693] env[61768]: DEBUG oslo_vmware.api [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228834, 'name': PowerOnVM_Task, 'duration_secs': 0.665543} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.313430] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 785.313721] env[61768]: INFO nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Took 9.61 seconds to spawn the instance on the hypervisor. [ 785.313721] env[61768]: DEBUG nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 785.315405] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247faabd-476a-431d-8385-7592cbefdfb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.355504] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228835, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100558} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.355996] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 785.356965] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-599152cd-71d2-4ba7-80d4-804c03092469 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.390286] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] fbb8c9db-e823-444a-84de-b231cf9b8a8a/fbb8c9db-e823-444a-84de-b231cf9b8a8a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 785.391828] env[61768]: DEBUG nova.network.neutron [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Successfully created port: 0e23d603-ab56-4c07-b562-335a39497bac {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.394436] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ae000558-5d35-4d95-ba24-50ed0cca8b3d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.420744] env[61768]: DEBUG oslo_vmware.api [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228836, 'name': PowerOffVM_Task, 'duration_secs': 0.251179} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.422349] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 785.422538] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 785.422948] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 785.422948] env[61768]: value = "task-1228839" [ 785.422948] env[61768]: _type = "Task" [ 785.422948] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.423183] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07398c98-e905-4f5f-aa85-3cd8dab01e6e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.435630] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228839, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.450836] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 785.451079] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 785.454927] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Deleting the datastore file [datastore1] 77319462-e447-405b-9269-82581effe005 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.454927] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aca52db2-1a27-4029-ab7c-45f57cc51b23 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.458622] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.459180] env[61768]: DEBUG oslo_concurrency.lockutils [req-a68cd3bc-a4df-47f2-890b-aaeeb9c8cea9 req-8d85286e-f271-4b4b-a591-2cd19380c20f service nova] Releasing lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.462426] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for the task: (returnval){ [ 785.462426] env[61768]: value = "task-1228841" [ 785.462426] env[61768]: _type = "Task" [ 785.462426] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.474143] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228841, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.479929] env[61768]: DEBUG nova.network.neutron [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Updated VIF entry in instance network info cache for port 61d36258-9772-43d5-aa74-55c86e1596a7. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.480526] env[61768]: DEBUG nova.network.neutron [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Updating instance_info_cache with network_info: [{"id": "61d36258-9772-43d5-aa74-55c86e1596a7", "address": "fa:16:3e:4c:2d:3b", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61d36258-97", "ovs_interfaceid": "61d36258-9772-43d5-aa74-55c86e1596a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.532390] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 785.532621] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 785.532942] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Deleting the datastore file [datastore1] a428f5e0-ade1-4aa6-af9d-0e33efcfec62 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.533533] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c3681aa8-9d62-45e4-bd1a-09570a71bc45 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.539886] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.545637] env[61768]: DEBUG oslo_vmware.api [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for the task: (returnval){ [ 785.545637] env[61768]: value = "task-1228842" [ 785.545637] env[61768]: _type = "Task" [ 785.545637] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.555040] env[61768]: DEBUG oslo_vmware.api [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228842, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.632405] env[61768]: DEBUG nova.scheduler.client.report [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.718808] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a2dec-08d9-66a5-4e0e-6299d3d44f80, 'name': SearchDatastore_Task, 'duration_secs': 0.051446} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.719788] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10f941de-7ecb-4be7-a740-abe472045490 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.726510] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 785.726510] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52502a1c-1f98-5982-e18a-cf6ac31f3999" [ 785.726510] env[61768]: _type = "Task" [ 785.726510] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.739254] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52502a1c-1f98-5982-e18a-cf6ac31f3999, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.835058] env[61768]: INFO nova.compute.manager [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Took 33.84 seconds to build instance. [ 785.936736] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228839, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.975384] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228841, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.983893] env[61768]: DEBUG oslo_concurrency.lockutils [req-282d1397-b3f2-4b7d-a6a4-a3a5c5e58aee req-9ce3e828-97b2-497d-9193-dd6c1b1d7431 service nova] Releasing lock "refresh_cache-3d96e47d-fc09-439e-bf0e-e60fc87ae958" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.034610] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.057371] env[61768]: DEBUG oslo_vmware.api [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Task: {'id': task-1228842, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.467559} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.057634] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.057898] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 786.058107] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 786.058321] env[61768]: INFO nova.compute.manager [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Took 1.19 seconds to destroy the instance on the hypervisor. [ 786.058579] env[61768]: DEBUG oslo.service.loopingcall [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.058815] env[61768]: DEBUG nova.compute.manager [-] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.058938] env[61768]: DEBUG nova.network.neutron [-] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 786.122785] env[61768]: DEBUG nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.142624] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.055s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.142624] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.860s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.149558] env[61768]: INFO nova.compute.claims [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.167702] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.167702] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.167964] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.168195] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.168399] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.168509] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.168956] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.169193] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.169396] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.169575] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.169769] env[61768]: DEBUG nova.virt.hardware [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.171781] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5d0b0f-d557-4b42-9e4a-704727ddf61b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.175918] env[61768]: INFO nova.scheduler.client.report [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Deleted allocations for instance d70040f1-e425-4aeb-b82d-ce808a02a645 [ 786.188469] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d1e750-4813-4611-813e-81c75b3585ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.241866] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52502a1c-1f98-5982-e18a-cf6ac31f3999, 'name': SearchDatastore_Task, 'duration_secs': 0.059696} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.242195] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.242467] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 3d96e47d-fc09-439e-bf0e-e60fc87ae958/3d96e47d-fc09-439e-bf0e-e60fc87ae958.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 786.242959] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5d31f94-bb1d-4246-9728-9942de170df5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.250805] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 786.250805] env[61768]: value = "task-1228843" [ 786.250805] env[61768]: _type = "Task" [ 786.250805] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.262173] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228843, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.338123] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fb49563-0877-483a-b4e2-ff865666ded4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.038s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.436968] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228839, 'name': ReconfigVM_Task, 'duration_secs': 0.901115} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.437392] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Reconfigured VM instance instance-00000040 to attach disk [datastore2] fbb8c9db-e823-444a-84de-b231cf9b8a8a/fbb8c9db-e823-444a-84de-b231cf9b8a8a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 786.438137] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7c0823dc-bb1d-462d-9b48-1bbe34d375aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.446587] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 786.446587] env[61768]: value = "task-1228844" [ 786.446587] env[61768]: _type = "Task" [ 786.446587] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.456181] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228844, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.473357] env[61768]: DEBUG oslo_vmware.api [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Task: {'id': task-1228841, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.548342} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.473681] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.473918] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 786.474134] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 786.474346] env[61768]: INFO nova.compute.manager [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] [instance: 77319462-e447-405b-9269-82581effe005] Took 2.24 seconds to destroy the instance on the hypervisor. [ 786.474616] env[61768]: DEBUG oslo.service.loopingcall [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.474850] env[61768]: DEBUG nova.compute.manager [-] [instance: 77319462-e447-405b-9269-82581effe005] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.474961] env[61768]: DEBUG nova.network.neutron [-] [instance: 77319462-e447-405b-9269-82581effe005] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 786.536270] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.559326] env[61768]: DEBUG nova.compute.manager [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.559759] env[61768]: DEBUG nova.compute.manager [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing instance network info cache due to event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 786.560200] env[61768]: DEBUG oslo_concurrency.lockutils [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.560490] env[61768]: DEBUG oslo_concurrency.lockutils [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.560846] env[61768]: DEBUG nova.network.neutron [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 786.687158] env[61768]: DEBUG oslo_concurrency.lockutils [None req-86e47f6d-9e15-43d9-86f6-32608d476f6c tempest-InstanceActionsV221TestJSON-1798251624 tempest-InstanceActionsV221TestJSON-1798251624-project-member] Lock "d70040f1-e425-4aeb-b82d-ce808a02a645" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.293s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.766331] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228843, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.841212] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 786.911873] env[61768]: DEBUG nova.network.neutron [-] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.957610] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228844, 'name': Rename_Task, 'duration_secs': 0.161746} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.958624] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 786.958930] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-835b29bf-6f98-4400-b0fc-0f7b5a4f183d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.967758] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 786.967758] env[61768]: value = "task-1228845" [ 786.967758] env[61768]: _type = "Task" [ 786.967758] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.979884] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228845, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.041260] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.270493] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228843, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.66811} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.270493] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 3d96e47d-fc09-439e-bf0e-e60fc87ae958/3d96e47d-fc09-439e-bf0e-e60fc87ae958.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 787.270493] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 787.270856] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-22abc451-7784-440c-bc16-893f1b096950 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.281272] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 787.281272] env[61768]: value = "task-1228846" [ 787.281272] env[61768]: _type = "Task" [ 787.281272] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.306507] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228846, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.364514] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.387708] env[61768]: DEBUG nova.network.neutron [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Successfully updated port: 0e23d603-ab56-4c07-b562-335a39497bac {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 787.414679] env[61768]: INFO nova.compute.manager [-] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Took 1.36 seconds to deallocate network for instance. [ 787.423181] env[61768]: DEBUG nova.network.neutron [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updated VIF entry in instance network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 787.423554] env[61768]: DEBUG nova.network.neutron [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.479996] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228845, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.545587] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.549397] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5576fc0-4afa-4766-9886-ce47ec685b02 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.557819] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a0816d-5e19-40a4-8f6c-73dc3287bbdb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.592644] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcaba790-6871-402f-bbcf-5bd37e35958e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.601830] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acecbff4-2de9-4191-8a74-dee401206d69 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.617841] env[61768]: DEBUG nova.compute.provider_tree [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.729436] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "37b729bf-ad5e-4e17-b11b-77b504c049cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.729622] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.729849] env[61768]: DEBUG nova.network.neutron [-] [instance: 77319462-e447-405b-9269-82581effe005] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.799134] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228846, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.262839} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.799442] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 787.800254] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f542a1-204c-4428-a2ad-eb55ede99fb3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.829573] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 3d96e47d-fc09-439e-bf0e-e60fc87ae958/3d96e47d-fc09-439e-bf0e-e60fc87ae958.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 787.829924] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2145112f-2562-497b-af6c-db1346eb372a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.853413] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 787.853413] env[61768]: value = "task-1228847" [ 787.853413] env[61768]: _type = "Task" [ 787.853413] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.862816] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228847, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.890764] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.890951] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.891129] env[61768]: DEBUG nova.network.neutron [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 787.926550] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.928296] env[61768]: DEBUG oslo_concurrency.lockutils [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.928651] env[61768]: DEBUG nova.compute.manager [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Received event network-vif-deleted-0416fa92-b8f1-4664-b2e1-df607b4200a3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.928913] env[61768]: INFO nova.compute.manager [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Neutron deleted interface 0416fa92-b8f1-4664-b2e1-df607b4200a3; detaching it from the instance and deleting it from the info cache [ 787.929179] env[61768]: DEBUG nova.network.neutron [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.989571] env[61768]: DEBUG oslo_vmware.api [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228845, 'name': PowerOnVM_Task, 'duration_secs': 0.963139} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.989899] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 787.990192] env[61768]: INFO nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Took 9.75 seconds to spawn the instance on the hypervisor. [ 787.990419] env[61768]: DEBUG nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 787.991731] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-523f531d-ab51-4479-afc8-7131fb5c921d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.039426] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.121134] env[61768]: DEBUG nova.scheduler.client.report [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.232501] env[61768]: INFO nova.compute.manager [-] [instance: 77319462-e447-405b-9269-82581effe005] Took 1.76 seconds to deallocate network for instance. [ 788.365862] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228847, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.438021] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4cd77274-a88d-4139-8f84-41661e62696e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.442680] env[61768]: DEBUG nova.network.neutron [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 788.448795] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1c8da8-86ff-423b-844c-3341fb450324 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.480071] env[61768]: DEBUG nova.compute.manager [req-48138084-0f60-4af0-9a18-ee178ef0ddc4 req-f49e26fd-2267-45fb-85c2-9bc858953a08 service nova] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Detach interface failed, port_id=0416fa92-b8f1-4664-b2e1-df607b4200a3, reason: Instance a428f5e0-ade1-4aa6-af9d-0e33efcfec62 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 788.516641] env[61768]: INFO nova.compute.manager [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Took 35.78 seconds to build instance. [ 788.541015] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.586790] env[61768]: DEBUG nova.compute.manager [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received event network-vif-deleted-5c1feaf9-e399-4171-90ae-e50db59e5274 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.587086] env[61768]: DEBUG nova.compute.manager [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Received event network-vif-plugged-0e23d603-ab56-4c07-b562-335a39497bac {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.587313] env[61768]: DEBUG oslo_concurrency.lockutils [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] Acquiring lock "4f429be7-c594-4a72-aebd-0b746219231c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.587524] env[61768]: DEBUG oslo_concurrency.lockutils [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] Lock "4f429be7-c594-4a72-aebd-0b746219231c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.587730] env[61768]: DEBUG oslo_concurrency.lockutils [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] Lock "4f429be7-c594-4a72-aebd-0b746219231c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.587887] env[61768]: DEBUG nova.compute.manager [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] No waiting events found dispatching network-vif-plugged-0e23d603-ab56-4c07-b562-335a39497bac {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 788.588087] env[61768]: WARNING nova.compute.manager [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Received unexpected event network-vif-plugged-0e23d603-ab56-4c07-b562-335a39497bac for instance with vm_state building and task_state spawning. [ 788.588260] env[61768]: DEBUG nova.compute.manager [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 77319462-e447-405b-9269-82581effe005] Received event network-vif-deleted-af1a7711-f21e-4462-93ef-a052655ce43c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.588426] env[61768]: DEBUG nova.compute.manager [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Received event network-changed-0e23d603-ab56-4c07-b562-335a39497bac {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.588576] env[61768]: DEBUG nova.compute.manager [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Refreshing instance network info cache due to event network-changed-0e23d603-ab56-4c07-b562-335a39497bac. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 788.588810] env[61768]: DEBUG oslo_concurrency.lockutils [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] Acquiring lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.626702] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.627253] env[61768]: DEBUG nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 788.630328] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.894s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.631796] env[61768]: INFO nova.compute.claims [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.698949] env[61768]: DEBUG nova.network.neutron [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance_info_cache with network_info: [{"id": "0e23d603-ab56-4c07-b562-335a39497bac", "address": "fa:16:3e:3d:a9:43", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e23d603-ab", "ovs_interfaceid": "0e23d603-ab56-4c07-b562-335a39497bac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.743459] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.866864] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228847, 'name': ReconfigVM_Task, 'duration_secs': 0.63927} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.867393] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 3d96e47d-fc09-439e-bf0e-e60fc87ae958/3d96e47d-fc09-439e-bf0e-e60fc87ae958.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 788.869114] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d07412c-643d-4b33-be58-eead0592b221 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.876725] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 788.876725] env[61768]: value = "task-1228848" [ 788.876725] env[61768]: _type = "Task" [ 788.876725] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.890710] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228848, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.019323] env[61768]: DEBUG oslo_concurrency.lockutils [None req-96f3c50c-3a0c-45d8-a322-a06b9f28ce8a tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.847s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.040728] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.136601] env[61768]: DEBUG nova.compute.utils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.142955] env[61768]: DEBUG nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.146474] env[61768]: DEBUG nova.network.neutron [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 789.205828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.206613] env[61768]: DEBUG nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Instance network_info: |[{"id": "0e23d603-ab56-4c07-b562-335a39497bac", "address": "fa:16:3e:3d:a9:43", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e23d603-ab", "ovs_interfaceid": "0e23d603-ab56-4c07-b562-335a39497bac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 789.207014] env[61768]: DEBUG oslo_concurrency.lockutils [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] Acquired lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.207313] env[61768]: DEBUG nova.network.neutron [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Refreshing network info cache for port 0e23d603-ab56-4c07-b562-335a39497bac {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 789.208621] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:a9:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0e23d603-ab56-4c07-b562-335a39497bac', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 789.219112] env[61768]: DEBUG oslo.service.loopingcall [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.223980] env[61768]: DEBUG nova.policy [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c39490b7563b4b20abbe27606b42f58d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f844c9d775474fb69d9ea04406cc1632', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.226489] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 789.229928] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59eb8cb5-a8b5-4559-a3ef-9b185503c196 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.256651] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 789.256651] env[61768]: value = "task-1228849" [ 789.256651] env[61768]: _type = "Task" [ 789.256651] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.265467] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228849, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.391212] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228848, 'name': Rename_Task, 'duration_secs': 0.276527} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.391212] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 789.391212] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-864e334b-cb53-45fc-b2a7-78ece8f627fc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.398510] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 789.398510] env[61768]: value = "task-1228850" [ 789.398510] env[61768]: _type = "Task" [ 789.398510] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.408596] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228850, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.522108] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 789.542021] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.580330] env[61768]: DEBUG nova.network.neutron [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updated VIF entry in instance network info cache for port 0e23d603-ab56-4c07-b562-335a39497bac. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 789.580706] env[61768]: DEBUG nova.network.neutron [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance_info_cache with network_info: [{"id": "0e23d603-ab56-4c07-b562-335a39497bac", "address": "fa:16:3e:3d:a9:43", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e23d603-ab", "ovs_interfaceid": "0e23d603-ab56-4c07-b562-335a39497bac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.647022] env[61768]: DEBUG nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 789.712050] env[61768]: DEBUG nova.network.neutron [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Successfully created port: 51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.767807] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228849, 'name': CreateVM_Task, 'duration_secs': 0.336335} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.767991] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 789.768723] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.768916] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.769473] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 789.772435] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d29f0f4-67e9-4319-aa2e-eaa77168687d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.779749] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 789.779749] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52293e2e-438b-a0e9-f91c-d9985ff14ad3" [ 789.779749] env[61768]: _type = "Task" [ 789.779749] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.790025] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52293e2e-438b-a0e9-f91c-d9985ff14ad3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.822062] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.822251] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.822504] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.822749] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.823801] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.825264] env[61768]: INFO nova.compute.manager [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Terminating instance [ 789.832124] env[61768]: DEBUG nova.compute.manager [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 789.833034] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 789.833257] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d61864-d5fa-46ee-9f4d-5e7af8f61edd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.843080] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 789.845939] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7dced925-d823-4579-9399-fab9d997e8f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.855463] env[61768]: DEBUG oslo_vmware.api [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 789.855463] env[61768]: value = "task-1228851" [ 789.855463] env[61768]: _type = "Task" [ 789.855463] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.868915] env[61768]: DEBUG oslo_vmware.api [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228851, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.913865] env[61768]: DEBUG oslo_vmware.api [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228850, 'name': PowerOnVM_Task, 'duration_secs': 0.467376} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.914205] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 789.914449] env[61768]: INFO nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Took 8.23 seconds to spawn the instance on the hypervisor. [ 789.914588] env[61768]: DEBUG nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 789.916187] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b826697-54fe-4fd0-8c9b-fb03ebefecea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.047238] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.049198] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350e500f-039c-4685-9e8d-bef5fe6fcc91 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.052450] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.058102] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea5b77f4-783d-49a3-98be-0e7dc92698fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.091468] env[61768]: DEBUG oslo_concurrency.lockutils [req-f010946c-4c3f-4b14-bb45-77162734b1f4 req-ce7ae1ed-47b9-4ab2-8552-ad53754635ba service nova] Releasing lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.092672] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73b96ef9-5518-4b3f-a113-92d956733204 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.101631] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66c106c9-1c25-41ac-be00-2430e3c0db2c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.117097] env[61768]: DEBUG nova.compute.provider_tree [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.291279] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52293e2e-438b-a0e9-f91c-d9985ff14ad3, 'name': SearchDatastore_Task, 'duration_secs': 0.023755} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.291609] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.291843] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 790.292102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.292262] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.292498] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 790.292728] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2a969fb4-4138-4d6b-b8f0-1d0f1c08ae2b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.304573] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 790.304975] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 790.306383] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf9149b8-0017-46f1-9172-9f53a32de44e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.313601] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 790.313601] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52693819-fbe6-6131-ebe6-eff1bc825d96" [ 790.313601] env[61768]: _type = "Task" [ 790.313601] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.328066] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52693819-fbe6-6131-ebe6-eff1bc825d96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.365296] env[61768]: DEBUG oslo_vmware.api [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228851, 'name': PowerOffVM_Task, 'duration_secs': 0.172517} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.365545] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 790.365731] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 790.366018] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b673ff58-6360-4cfe-be6c-0b09007c0a7d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.440346] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 790.440565] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 790.441122] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Deleting the datastore file [datastore2] fbb8c9db-e823-444a-84de-b231cf9b8a8a {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 790.441122] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5770f2f7-3563-42aa-a71f-939e476c5964 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.443456] env[61768]: INFO nova.compute.manager [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Took 32.55 seconds to build instance. [ 790.449954] env[61768]: DEBUG oslo_vmware.api [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for the task: (returnval){ [ 790.449954] env[61768]: value = "task-1228853" [ 790.449954] env[61768]: _type = "Task" [ 790.449954] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.459924] env[61768]: DEBUG oslo_vmware.api [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228853, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.543421] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.620642] env[61768]: DEBUG nova.scheduler.client.report [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.661267] env[61768]: DEBUG nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 790.696788] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 790.697421] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 790.697864] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 790.701324] env[61768]: DEBUG nova.virt.hardware [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 790.701324] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1004acc4-b353-4985-bc63-1a07e60e1156 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.711576] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a4a1713-ff3e-4f74-8f02-7ae68f71d06e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.826806] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52693819-fbe6-6131-ebe6-eff1bc825d96, 'name': SearchDatastore_Task, 'duration_secs': 0.018685} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.827711] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eaa95b1d-6a4c-4015-b12f-a95e4ed48d24 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.834126] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 790.834126] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5243dc16-b5c9-efd8-4411-069be80c810d" [ 790.834126] env[61768]: _type = "Task" [ 790.834126] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.843694] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5243dc16-b5c9-efd8-4411-069be80c810d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.945287] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6b837ef2-6800-40ca-a89b-d9705d38059f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.556s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.960235] env[61768]: DEBUG oslo_vmware.api [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Task: {'id': task-1228853, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.236736} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.960511] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 790.960704] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 790.960882] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 790.961073] env[61768]: INFO nova.compute.manager [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Took 1.13 seconds to destroy the instance on the hypervisor. [ 790.961324] env[61768]: DEBUG oslo.service.loopingcall [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.962128] env[61768]: DEBUG nova.compute.manager [-] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.962225] env[61768]: DEBUG nova.network.neutron [-] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 791.044385] env[61768]: DEBUG oslo_vmware.api [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228837, 'name': ReconfigVM_Task, 'duration_secs': 5.83712} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.044760] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.044911] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Reconfigured VM to detach interface {{(pid=61768) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 791.129381] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.129998] env[61768]: DEBUG nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.132975] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.933s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.136417] env[61768]: INFO nova.compute.claims [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.348452] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5243dc16-b5c9-efd8-4411-069be80c810d, 'name': SearchDatastore_Task, 'duration_secs': 0.009964} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.348766] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.349167] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 4f429be7-c594-4a72-aebd-0b746219231c/4f429be7-c594-4a72-aebd-0b746219231c.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 791.349479] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1d92ea3-b2a9-4e15-8121-43f8b8b54ad9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.356281] env[61768]: DEBUG nova.compute.manager [req-982518f7-22df-4295-ac1a-7dc265a84585 req-ef8d818a-097c-431f-a7ea-1460f07153b2 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Received event network-vif-plugged-51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.356561] env[61768]: DEBUG oslo_concurrency.lockutils [req-982518f7-22df-4295-ac1a-7dc265a84585 req-ef8d818a-097c-431f-a7ea-1460f07153b2 service nova] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.356820] env[61768]: DEBUG oslo_concurrency.lockutils [req-982518f7-22df-4295-ac1a-7dc265a84585 req-ef8d818a-097c-431f-a7ea-1460f07153b2 service nova] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.357097] env[61768]: DEBUG oslo_concurrency.lockutils [req-982518f7-22df-4295-ac1a-7dc265a84585 req-ef8d818a-097c-431f-a7ea-1460f07153b2 service nova] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.357272] env[61768]: DEBUG nova.compute.manager [req-982518f7-22df-4295-ac1a-7dc265a84585 req-ef8d818a-097c-431f-a7ea-1460f07153b2 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] No waiting events found dispatching network-vif-plugged-51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 791.357488] env[61768]: WARNING nova.compute.manager [req-982518f7-22df-4295-ac1a-7dc265a84585 req-ef8d818a-097c-431f-a7ea-1460f07153b2 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Received unexpected event network-vif-plugged-51b2319f-ceb1-41db-8019-1d918e70080d for instance with vm_state building and task_state spawning. [ 791.359443] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 791.359443] env[61768]: value = "task-1228854" [ 791.359443] env[61768]: _type = "Task" [ 791.359443] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.369378] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.448178] env[61768]: DEBUG nova.compute.manager [None req-f1339579-c392-4725-9b33-c0e6bf0be676 tempest-ServersListShow296Test-1960977358 tempest-ServersListShow296Test-1960977358-project-member] [instance: 54d8c044-3cc6-4e0c-a809-cc8351b80b27] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.539583] env[61768]: DEBUG nova.compute.manager [req-961ebc9f-965a-46af-85cd-e5b6714de1df req-2134df94-7d8e-470b-ba4a-774d909d95a3 service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Received event network-vif-deleted-3bc689b6-7278-4741-8053-b98ed2db0265 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.539887] env[61768]: INFO nova.compute.manager [req-961ebc9f-965a-46af-85cd-e5b6714de1df req-2134df94-7d8e-470b-ba4a-774d909d95a3 service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Neutron deleted interface 3bc689b6-7278-4741-8053-b98ed2db0265; detaching it from the instance and deleting it from the info cache [ 791.540164] env[61768]: DEBUG nova.network.neutron [req-961ebc9f-965a-46af-85cd-e5b6714de1df req-2134df94-7d8e-470b-ba4a-774d909d95a3 service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.646171] env[61768]: DEBUG nova.compute.utils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.649370] env[61768]: DEBUG nova.network.neutron [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Successfully updated port: 51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 791.651830] env[61768]: DEBUG nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.652865] env[61768]: DEBUG nova.network.neutron [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 791.703580] env[61768]: DEBUG nova.policy [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62522bd3731a47bc928a0c4038c447d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06ce9d5af7e740ff882439e4486a0aad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 791.870900] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228854, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.957339] env[61768]: DEBUG nova.compute.manager [None req-f1339579-c392-4725-9b33-c0e6bf0be676 tempest-ServersListShow296Test-1960977358 tempest-ServersListShow296Test-1960977358-project-member] [instance: 54d8c044-3cc6-4e0c-a809-cc8351b80b27] Instance disappeared before build. {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 792.008084] env[61768]: DEBUG nova.network.neutron [-] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.046417] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f05402a-32a1-4283-90a6-2aaead599605 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.057036] env[61768]: DEBUG nova.compute.manager [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.059807] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dccfc6c-518f-4e2b-898d-f830d4dd7dd9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.067389] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de72a01a-51a8-4c96-bd80-799d0717dcdf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.081475] env[61768]: DEBUG nova.network.neutron [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Successfully created port: c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.107707] env[61768]: DEBUG nova.compute.manager [req-961ebc9f-965a-46af-85cd-e5b6714de1df req-2134df94-7d8e-470b-ba4a-774d909d95a3 service nova] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Detach interface failed, port_id=3bc689b6-7278-4741-8053-b98ed2db0265, reason: Instance fbb8c9db-e823-444a-84de-b231cf9b8a8a could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 792.158035] env[61768]: DEBUG nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.162026] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.163786] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.163786] env[61768]: DEBUG nova.network.neutron [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 792.376805] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.870448} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.377183] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 4f429be7-c594-4a72-aebd-0b746219231c/4f429be7-c594-4a72-aebd-0b746219231c.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 792.377452] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 792.377783] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-518ef189-2258-4dc7-bb3e-2f1bef9bac42 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.387271] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 792.387271] env[61768]: value = "task-1228855" [ 792.387271] env[61768]: _type = "Task" [ 792.387271] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.402797] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228855, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.482306] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1339579-c392-4725-9b33-c0e6bf0be676 tempest-ServersListShow296Test-1960977358 tempest-ServersListShow296Test-1960977358-project-member] Lock "54d8c044-3cc6-4e0c-a809-cc8351b80b27" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 9.812s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.511344] env[61768]: INFO nova.compute.manager [-] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Took 1.55 seconds to deallocate network for instance. [ 792.531727] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-522a2b8a-df10-473e-acec-224946ca5587 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.543308] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb3348c-dcb6-4430-978a-52124bd044ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.581286] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eeb77d9-b411-4650-ae50-944709d4efc7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.590421] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a98d1e-bebb-4c13-ab0b-9f081e2da258 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.598500] env[61768]: INFO nova.compute.manager [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] instance snapshotting [ 792.602617] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167c685b-db75-48ce-adf3-026c6cedb1b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.614668] env[61768]: DEBUG nova.compute.provider_tree [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.635956] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.636100] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.640019] env[61768]: DEBUG nova.network.neutron [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 792.640019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f355402e-f2f4-4d26-a34e-4a02b1b32bff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.711297] env[61768]: DEBUG nova.network.neutron [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 792.900624] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228855, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077034} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.900915] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.905019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f4a2c6-3333-4ced-8cc5-6611d345e71e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.928182] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 4f429be7-c594-4a72-aebd-0b746219231c/4f429be7-c594-4a72-aebd-0b746219231c.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.928182] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-927264aa-107c-4533-9a3d-ab0e6ec48401 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.949358] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 792.949358] env[61768]: value = "task-1228856" [ 792.949358] env[61768]: _type = "Task" [ 792.949358] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.950331] env[61768]: DEBUG nova.network.neutron [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating instance_info_cache with network_info: [{"id": "51b2319f-ceb1-41db-8019-1d918e70080d", "address": "fa:16:3e:7d:47:ab", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b2319f-ce", "ovs_interfaceid": "51b2319f-ceb1-41db-8019-1d918e70080d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.962601] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228856, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.986999] env[61768]: DEBUG nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 793.022817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.117646] env[61768]: DEBUG nova.scheduler.client.report [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.152621] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 793.153234] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-20c8ea28-5d14-44b6-b952-02541858c572 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.162743] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 793.162743] env[61768]: value = "task-1228857" [ 793.162743] env[61768]: _type = "Task" [ 793.162743] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.173533] env[61768]: DEBUG nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.189176] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.189176] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "12a543c2-1081-49d7-800b-07f0a2516904" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.189649] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.189649] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.189857] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "12a543c2-1081-49d7-800b-07f0a2516904-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.191509] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228857, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.193137] env[61768]: INFO nova.compute.manager [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Terminating instance [ 793.195829] env[61768]: DEBUG nova.compute.manager [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.196065] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 793.196906] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40914e76-df0a-495c-b728-c248bf2a331f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.209491] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 793.214296] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78a565ae-a686-44a5-ba06-443ec1bb23b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.224027] env[61768]: DEBUG nova.virt.hardware [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.226061] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b451090b-a8d6-4c48-82ad-b0128fda5476 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.231638] env[61768]: DEBUG oslo_vmware.api [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 793.231638] env[61768]: value = "task-1228858" [ 793.231638] env[61768]: _type = "Task" [ 793.231638] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.240942] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e6af29-ddca-4d04-8d8e-7d8dfcf16c2a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.249024] env[61768]: DEBUG oslo_vmware.api [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228858, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.386285] env[61768]: DEBUG nova.compute.manager [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Received event network-changed-51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.386600] env[61768]: DEBUG nova.compute.manager [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Refreshing instance network info cache due to event network-changed-51b2319f-ceb1-41db-8019-1d918e70080d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.386708] env[61768]: DEBUG oslo_concurrency.lockutils [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] Acquiring lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.457676] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.458075] env[61768]: DEBUG nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Instance network_info: |[{"id": "51b2319f-ceb1-41db-8019-1d918e70080d", "address": "fa:16:3e:7d:47:ab", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b2319f-ce", "ovs_interfaceid": "51b2319f-ceb1-41db-8019-1d918e70080d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 793.458419] env[61768]: DEBUG oslo_concurrency.lockutils [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] Acquired lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.458622] env[61768]: DEBUG nova.network.neutron [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Refreshing network info cache for port 51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 793.459980] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:47:ab', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc845e3-654b-43c6-acea-dde1084f0ad0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51b2319f-ceb1-41db-8019-1d918e70080d', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 793.469322] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Creating folder: Project (f844c9d775474fb69d9ea04406cc1632). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 793.470392] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e4e8807d-b317-40ea-8a6c-29c012911345 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.477633] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228856, 'name': ReconfigVM_Task, 'duration_secs': 0.484575} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.479358] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 4f429be7-c594-4a72-aebd-0b746219231c/4f429be7-c594-4a72-aebd-0b746219231c.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.480646] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f45f5c8-e6b2-42b7-9e6e-6ec7c2553375 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.489628] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 793.489628] env[61768]: value = "task-1228860" [ 793.489628] env[61768]: _type = "Task" [ 793.489628] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.503384] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Created folder: Project (f844c9d775474fb69d9ea04406cc1632) in parent group-v265360. [ 793.503604] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Creating folder: Instances. Parent ref: group-v265444. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 793.504748] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-251fc192-89da-4355-bc14-8e4edc4dfc58 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.511424] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228860, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.521661] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.525539] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Created folder: Instances in parent group-v265444. [ 793.525939] env[61768]: DEBUG oslo.service.loopingcall [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.526224] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 793.526508] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c93747af-6ce6-4673-a199-55a3def782d6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.554414] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 793.554414] env[61768]: value = "task-1228862" [ 793.554414] env[61768]: _type = "Task" [ 793.554414] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.567447] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228862, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.607975] env[61768]: DEBUG nova.compute.manager [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-vif-deleted-a0589e32-427b-4fea-a7f6-6de811fcebdb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.608235] env[61768]: INFO nova.compute.manager [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Neutron deleted interface a0589e32-427b-4fea-a7f6-6de811fcebdb; detaching it from the instance and deleting it from the info cache [ 793.608531] env[61768]: DEBUG nova.network.neutron [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "cb8850cf-e632-419a-86f0-4c532e443213", "address": "fa:16:3e:c8:6d:50", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcb8850cf-e6", "ovs_interfaceid": "cb8850cf-e632-419a-86f0-4c532e443213", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.623959] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.491s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.624330] env[61768]: DEBUG nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.637719] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.748s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.644023] env[61768]: INFO nova.compute.claims [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 793.676159] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228857, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.743322] env[61768]: DEBUG oslo_vmware.api [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228858, 'name': PowerOffVM_Task, 'duration_secs': 0.24341} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.743744] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 793.744083] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 793.744471] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f56c8b7e-8ec6-45f8-81c0-df1dbaa62f47 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.832954] env[61768]: INFO nova.network.neutron [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Port a0589e32-427b-4fea-a7f6-6de811fcebdb from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 793.832954] env[61768]: INFO nova.network.neutron [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Port cb8850cf-e632-419a-86f0-4c532e443213 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 793.832954] env[61768]: DEBUG nova.network.neutron [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.871076] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 793.871076] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 793.871287] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleting the datastore file [datastore2] 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.872585] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1780fb08-458a-4918-8eca-b406bcb98600 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.878057] env[61768]: DEBUG nova.network.neutron [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Successfully updated port: c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 793.885161] env[61768]: DEBUG oslo_vmware.api [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 793.885161] env[61768]: value = "task-1228864" [ 793.885161] env[61768]: _type = "Task" [ 793.885161] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.898729] env[61768]: DEBUG oslo_vmware.api [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228864, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.002024] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228860, 'name': Rename_Task, 'duration_secs': 0.238182} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.002406] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 794.002923] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2832337c-50c7-4d91-8ab0-5a77a93dd077 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.011825] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 794.011825] env[61768]: value = "task-1228865" [ 794.011825] env[61768]: _type = "Task" [ 794.011825] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.022513] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228865, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.067052] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228862, 'name': CreateVM_Task, 'duration_secs': 0.453329} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.067238] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 794.067971] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.068170] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.068505] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 794.068782] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73791dee-e06a-4e90-a1bb-20be55b6861a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.075360] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 794.075360] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522ddd78-ab59-9df6-8b59-28130f2ce6dd" [ 794.075360] env[61768]: _type = "Task" [ 794.075360] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.087711] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522ddd78-ab59-9df6-8b59-28130f2ce6dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.111671] env[61768]: DEBUG oslo_concurrency.lockutils [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] Acquiring lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.137362] env[61768]: DEBUG nova.compute.utils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.138899] env[61768]: DEBUG nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.139165] env[61768]: DEBUG nova.network.neutron [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 794.176080] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228857, 'name': CreateSnapshot_Task, 'duration_secs': 0.845102} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.176437] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 794.177303] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-145d5b31-4e1c-429e-a17c-1a09c9c6449a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.206765] env[61768]: DEBUG nova.policy [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '33db714fc1044225ac0286708a0bb7c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa6da53b22194f598ac23dae71b24c07', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.335589] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.371655] env[61768]: DEBUG nova.network.neutron [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updated VIF entry in instance network info cache for port 51b2319f-ceb1-41db-8019-1d918e70080d. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 794.372043] env[61768]: DEBUG nova.network.neutron [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating instance_info_cache with network_info: [{"id": "51b2319f-ceb1-41db-8019-1d918e70080d", "address": "fa:16:3e:7d:47:ab", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b2319f-ce", "ovs_interfaceid": "51b2319f-ceb1-41db-8019-1d918e70080d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.385127] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "refresh_cache-a36b542c-334f-46b1-9030-ee5c9f8c55c1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.385127] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "refresh_cache-a36b542c-334f-46b1-9030-ee5c9f8c55c1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.385127] env[61768]: DEBUG nova.network.neutron [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 794.398669] env[61768]: DEBUG oslo_vmware.api [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1228864, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.20977} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.398669] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.399043] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 794.399043] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 794.399176] env[61768]: INFO nova.compute.manager [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Took 1.20 seconds to destroy the instance on the hypervisor. [ 794.399857] env[61768]: DEBUG oslo.service.loopingcall [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.400217] env[61768]: DEBUG nova.compute.manager [-] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.400324] env[61768]: DEBUG nova.network.neutron [-] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 794.479311] env[61768]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port cb8850cf-e632-419a-86f0-4c532e443213 could not be found.", "detail": ""}} {{(pid=61768) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 794.479572] env[61768]: DEBUG nova.network.neutron [-] Unable to show port cb8850cf-e632-419a-86f0-4c532e443213 as it no longer exists. {{(pid=61768) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 794.488385] env[61768]: DEBUG nova.network.neutron [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Successfully created port: 5405ba6b-c888-48a0-9af2-4024e9af563e {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.523957] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228865, 'name': PowerOnVM_Task} progress is 74%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.590014] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522ddd78-ab59-9df6-8b59-28130f2ce6dd, 'name': SearchDatastore_Task, 'duration_secs': 0.013485} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.590352] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.590352] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 794.590805] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 794.590974] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.591229] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 794.591551] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a53fa4e6-11ac-442c-9a15-d110e677e822 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.602822] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 794.607021] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 794.607021] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27369e18-d61a-42cb-815c-5b725f2ff8c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.611364] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 794.611364] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525b55d7-6021-e682-54a5-fe0076be8fbd" [ 794.611364] env[61768]: _type = "Task" [ 794.611364] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.621935] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525b55d7-6021-e682-54a5-fe0076be8fbd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.642753] env[61768]: DEBUG nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.698433] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 794.699812] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ce93353b-9085-4887-9d80-ed38d90d0469 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.715555] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 794.715555] env[61768]: value = "task-1228866" [ 794.715555] env[61768]: _type = "Task" [ 794.715555] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.728447] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228866, 'name': CloneVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.839699] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a0336dfa-12bf-47af-8326-66dd8e8d4c0d tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-12a543c2-1081-49d7-800b-07f0a2516904-a0589e32-427b-4fea-a7f6-6de811fcebdb" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.437s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.875599] env[61768]: DEBUG oslo_concurrency.lockutils [req-833cfa44-f30b-46a6-ad13-46f3d300be49 req-aa073a53-1675-4c2c-a865-3ccc40a54c53 service nova] Releasing lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.937654] env[61768]: DEBUG nova.network.neutron [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 795.033927] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228865, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.036168] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-766e2f6b-9887-48c5-a99e-208f0b9bb661 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.045396] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb18d967-9cc1-42d7-b899-451ad6caa0fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.083663] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80544b61-13d3-4f17-8f12-7be2f5f441e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.094920] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f195c4b9-49ec-4e21-9ab6-9c4fd7c0f5fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.112696] env[61768]: DEBUG nova.compute.provider_tree [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.122871] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525b55d7-6021-e682-54a5-fe0076be8fbd, 'name': SearchDatastore_Task, 'duration_secs': 0.010171} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.124892] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f3843a5-c902-4be8-80dc-210da4719e8e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.132982] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 795.132982] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bccaf4-b130-8e00-b2b7-bc62749c191c" [ 795.132982] env[61768]: _type = "Task" [ 795.132982] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.146313] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bccaf4-b130-8e00-b2b7-bc62749c191c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.194697] env[61768]: DEBUG nova.network.neutron [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Updating instance_info_cache with network_info: [{"id": "c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10", "address": "fa:16:3e:fe:7b:a8", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc87ca0a3-89", "ovs_interfaceid": "c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.228502] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228866, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.528877] env[61768]: DEBUG oslo_vmware.api [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228865, 'name': PowerOnVM_Task, 'duration_secs': 1.232609} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.528877] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 795.528877] env[61768]: INFO nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Took 9.40 seconds to spawn the instance on the hypervisor. [ 795.528877] env[61768]: DEBUG nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 795.528877] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867bba28-da2e-4c40-91fc-ac559c52f630 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.567715] env[61768]: DEBUG nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Received event network-vif-plugged-c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.567715] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] Acquiring lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.568247] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.568309] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.568470] env[61768]: DEBUG nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] No waiting events found dispatching network-vif-plugged-c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 795.568675] env[61768]: WARNING nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Received unexpected event network-vif-plugged-c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 for instance with vm_state building and task_state spawning. [ 795.568886] env[61768]: DEBUG nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Received event network-changed-c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.569078] env[61768]: DEBUG nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Refreshing instance network info cache due to event network-changed-c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.569262] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] Acquiring lock "refresh_cache-a36b542c-334f-46b1-9030-ee5c9f8c55c1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.617042] env[61768]: DEBUG nova.scheduler.client.report [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.648818] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bccaf4-b130-8e00-b2b7-bc62749c191c, 'name': SearchDatastore_Task, 'duration_secs': 0.014389} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.649157] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.649486] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 7697ecda-ef30-447d-a92a-3f5cb3cc9118/7697ecda-ef30-447d-a92a-3f5cb3cc9118.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 795.649778] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f0d0e8d-4e88-415d-b4c6-3a6f3d77fa2c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.653031] env[61768]: DEBUG nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.663581] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 795.663581] env[61768]: value = "task-1228867" [ 795.663581] env[61768]: _type = "Task" [ 795.663581] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.671361] env[61768]: DEBUG nova.network.neutron [-] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.676065] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228867, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.701487] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "refresh_cache-a36b542c-334f-46b1-9030-ee5c9f8c55c1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.701847] env[61768]: DEBUG nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Instance network_info: |[{"id": "c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10", "address": "fa:16:3e:fe:7b:a8", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc87ca0a3-89", "ovs_interfaceid": "c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 795.702582] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] Acquired lock "refresh_cache-a36b542c-334f-46b1-9030-ee5c9f8c55c1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.702788] env[61768]: DEBUG nova.network.neutron [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Refreshing network info cache for port c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 795.704122] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fe:7b:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3291573-fad8-48cc-a965-c3554e7cee4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.713046] env[61768]: DEBUG oslo.service.loopingcall [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.717779] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 795.717779] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e2203784-568a-464a-93ad-72852949cd57 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.751330] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228866, 'name': CloneVM_Task} progress is 95%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.754280] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.755223] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.755223] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.755223] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.755467] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.755526] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.755760] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.755947] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.756151] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.756324] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.756500] env[61768]: DEBUG nova.virt.hardware [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.760057] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d5d5552-83e9-4849-b538-0957263db958 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.762390] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.762390] env[61768]: value = "task-1228868" [ 795.762390] env[61768]: _type = "Task" [ 795.762390] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.771842] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-885d08a4-c8d6-4733-8f63-893302614bcb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.780983] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228868, 'name': CreateVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.841855] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "8e0e13a7-eac5-4176-8536-2906d13c390e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.842217] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.051783] env[61768]: INFO nova.compute.manager [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Took 35.43 seconds to build instance. [ 796.123632] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.124644] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 796.131929] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.425s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.132348] env[61768]: DEBUG nova.objects.instance [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'resources' on Instance uuid 76c282d1-ddfe-46dc-aa7b-225708443379 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.180970] env[61768]: INFO nova.compute.manager [-] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Took 1.78 seconds to deallocate network for instance. [ 796.183490] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228867, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.216031] env[61768]: DEBUG nova.network.neutron [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Updated VIF entry in instance network info cache for port c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 796.217230] env[61768]: DEBUG nova.network.neutron [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Updating instance_info_cache with network_info: [{"id": "c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10", "address": "fa:16:3e:fe:7b:a8", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc87ca0a3-89", "ovs_interfaceid": "c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.234487] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228866, 'name': CloneVM_Task, 'duration_secs': 1.202329} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.234806] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Created linked-clone VM from snapshot [ 796.235694] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4346bd25-4ba9-41f6-93fa-49be7613b7cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.247190] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Uploading image 6c2ff2ef-4663-4449-a0eb-4863e7b27db2 {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 796.275477] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 796.275477] env[61768]: value = "vm-265448" [ 796.275477] env[61768]: _type = "VirtualMachine" [ 796.275477] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 796.277784] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-fc1736c1-fb8e-4368-913e-ac50926176df {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.285023] env[61768]: DEBUG nova.compute.manager [req-2f7b1660-3742-43fc-bdac-035aa956fbfe req-87f93f5a-ee09-4e09-90b1-25544f9d8643 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Received event network-vif-plugged-5405ba6b-c888-48a0-9af2-4024e9af563e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.285564] env[61768]: DEBUG oslo_concurrency.lockutils [req-2f7b1660-3742-43fc-bdac-035aa956fbfe req-87f93f5a-ee09-4e09-90b1-25544f9d8643 service nova] Acquiring lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.286455] env[61768]: DEBUG oslo_concurrency.lockutils [req-2f7b1660-3742-43fc-bdac-035aa956fbfe req-87f93f5a-ee09-4e09-90b1-25544f9d8643 service nova] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.286455] env[61768]: DEBUG oslo_concurrency.lockutils [req-2f7b1660-3742-43fc-bdac-035aa956fbfe req-87f93f5a-ee09-4e09-90b1-25544f9d8643 service nova] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.287240] env[61768]: DEBUG nova.compute.manager [req-2f7b1660-3742-43fc-bdac-035aa956fbfe req-87f93f5a-ee09-4e09-90b1-25544f9d8643 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] No waiting events found dispatching network-vif-plugged-5405ba6b-c888-48a0-9af2-4024e9af563e {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 796.287240] env[61768]: WARNING nova.compute.manager [req-2f7b1660-3742-43fc-bdac-035aa956fbfe req-87f93f5a-ee09-4e09-90b1-25544f9d8643 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Received unexpected event network-vif-plugged-5405ba6b-c888-48a0-9af2-4024e9af563e for instance with vm_state building and task_state spawning. [ 796.293959] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228868, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.299982] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lease: (returnval){ [ 796.299982] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527c3135-6b6b-4449-2765-98dac96230a9" [ 796.299982] env[61768]: _type = "HttpNfcLease" [ 796.299982] env[61768]: } obtained for exporting VM: (result){ [ 796.299982] env[61768]: value = "vm-265448" [ 796.299982] env[61768]: _type = "VirtualMachine" [ 796.299982] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 796.299982] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the lease: (returnval){ [ 796.299982] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527c3135-6b6b-4449-2765-98dac96230a9" [ 796.299982] env[61768]: _type = "HttpNfcLease" [ 796.299982] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 796.311296] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 796.311296] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527c3135-6b6b-4449-2765-98dac96230a9" [ 796.311296] env[61768]: _type = "HttpNfcLease" [ 796.311296] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 796.392772] env[61768]: DEBUG nova.network.neutron [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Successfully updated port: 5405ba6b-c888-48a0-9af2-4024e9af563e {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 796.559496] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0fa5d181-a3e9-4c0a-8de3-30c90f8ac9be tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.278s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.636350] env[61768]: DEBUG nova.compute.utils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.641799] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.641799] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.680564] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228867, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.629203} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.682047] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 7697ecda-ef30-447d-a92a-3f5cb3cc9118/7697ecda-ef30-447d-a92a-3f5cb3cc9118.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 796.682047] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 796.682047] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a329dec3-d62a-4645-ad91-58da9be1005a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.686518] env[61768]: DEBUG nova.policy [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5721b9005cf8496995615cfc21da01a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '851a60883ea34523abf3a0ce81e6d40f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.691732] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.691941] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 796.691941] env[61768]: value = "task-1228870" [ 796.691941] env[61768]: _type = "Task" [ 796.691941] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.704708] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.722227] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] Releasing lock "refresh_cache-a36b542c-334f-46b1-9030-ee5c9f8c55c1" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.722502] env[61768]: DEBUG nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-vif-deleted-3a76322f-8de0-44f9-9b42-ae3f6c592d3f {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.723041] env[61768]: INFO nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Neutron deleted interface 3a76322f-8de0-44f9-9b42-ae3f6c592d3f; detaching it from the instance and deleting it from the info cache [ 796.723041] env[61768]: DEBUG nova.network.neutron [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.779143] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228868, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.814753] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 796.814753] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527c3135-6b6b-4449-2765-98dac96230a9" [ 796.814753] env[61768]: _type = "HttpNfcLease" [ 796.814753] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 796.814753] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 796.814753] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527c3135-6b6b-4449-2765-98dac96230a9" [ 796.814753] env[61768]: _type = "HttpNfcLease" [ 796.814753] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 796.814753] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61a0fc49-5c04-4ea7-b937-1031c566d71b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.823471] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52761ef9-9a27-a2d0-1c84-993bba1d4b81/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 796.826467] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52761ef9-9a27-a2d0-1c84-993bba1d4b81/disk-0.vmdk for reading. {{(pid=61768) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 796.898121] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "refresh_cache-23cc444d-ff15-441b-8a0c-2ded49fd3cf0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.898121] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "refresh_cache-23cc444d-ff15-441b-8a0c-2ded49fd3cf0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.898121] env[61768]: DEBUG nova.network.neutron [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 796.982487] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-0f00ec40-5b6a-48a7-af0a-803a89a128ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.055372] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcccfee6-5253-40cb-a121-4ab9a7d569ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.063942] env[61768]: DEBUG nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.067588] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7078019c-a25d-4d8b-ad79-0a3798299fa5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.110795] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Successfully created port: 2889bdd9-2069-495a-b80a-b20bbec5ef00 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.113221] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a960264-2699-4202-b580-9637090e64a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.126861] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef7bc9a-07a6-4350-8863-d18234d659f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.141293] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.144531] env[61768]: DEBUG nova.compute.provider_tree [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.206483] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124523} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.206483] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 797.207250] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbbfa8e8-b0bf-4b40-b574-58d9347535cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.236985] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Reconfiguring VM instance instance-00000043 to attach disk [datastore1] 7697ecda-ef30-447d-a92a-3f5cb3cc9118/7697ecda-ef30-447d-a92a-3f5cb3cc9118.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.237526] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79dc45a7-05d8-4034-a1cb-4a9f150af8b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.240694] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-88fb7da7-adc3-4656-9f13-6e3c0c5d6d70 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.272699] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcadb0f-4641-44cc-80df-a958653dbb89 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.287346] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 797.287346] env[61768]: value = "task-1228871" [ 797.287346] env[61768]: _type = "Task" [ 797.287346] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.295650] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228868, 'name': CreateVM_Task, 'duration_secs': 1.439623} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.296380] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 797.297227] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.297498] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.297977] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 797.298314] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-754ec907-82b9-4df8-8735-094636be96f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.312199] env[61768]: DEBUG nova.compute.manager [req-ed901ef4-3138-4063-ab54-178fc90a9f72 req-9681957a-d16b-4b1c-aa5c-8b3f78ac2450 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Detach interface failed, port_id=3a76322f-8de0-44f9-9b42-ae3f6c592d3f, reason: Instance 12a543c2-1081-49d7-800b-07f0a2516904 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 797.312899] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228871, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.317424] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 797.317424] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c34dfc-1002-c405-6e03-f400752408a8" [ 797.317424] env[61768]: _type = "Task" [ 797.317424] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.326885] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c34dfc-1002-c405-6e03-f400752408a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.458741] env[61768]: DEBUG nova.network.neutron [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 797.596779] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.651662] env[61768]: DEBUG nova.scheduler.client.report [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.741741] env[61768]: DEBUG nova.network.neutron [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Updating instance_info_cache with network_info: [{"id": "5405ba6b-c888-48a0-9af2-4024e9af563e", "address": "fa:16:3e:1c:2a:4f", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5405ba6b-c8", "ovs_interfaceid": "5405ba6b-c888-48a0-9af2-4024e9af563e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.801496] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228871, 'name': ReconfigVM_Task, 'duration_secs': 0.307258} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.802422] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Reconfigured VM instance instance-00000043 to attach disk [datastore1] 7697ecda-ef30-447d-a92a-3f5cb3cc9118/7697ecda-ef30-447d-a92a-3f5cb3cc9118.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.803551] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5ec5174e-6264-4060-bae6-2b8a3827ba0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.816020] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 797.816020] env[61768]: value = "task-1228872" [ 797.816020] env[61768]: _type = "Task" [ 797.816020] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.830965] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c34dfc-1002-c405-6e03-f400752408a8, 'name': SearchDatastore_Task, 'duration_secs': 0.010764} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.843196] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.847024] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.847024] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.847024] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.847024] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.847024] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228872, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.847024] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-524880c3-1f89-415e-a151-92ac34dfb766 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.861629] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.862810] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 797.864374] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25b8d2c7-8283-4dd2-a14c-70b7136c64d1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.872720] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 797.872720] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d0e31-d6f0-f029-d0f8-e4eed1ffd567" [ 797.872720] env[61768]: _type = "Task" [ 797.872720] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.883890] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d0e31-d6f0-f029-d0f8-e4eed1ffd567, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.157976] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.161194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.029s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.164054] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.231s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.165446] env[61768]: INFO nova.compute.claims [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.185523] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.185961] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.186212] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.186689] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.186978] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.187228] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.187490] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.187714] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.187965] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.188247] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.188660] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.189751] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab86e3ad-2cdf-4888-8838-b810c0031059 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.193441] env[61768]: INFO nova.scheduler.client.report [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted allocations for instance 76c282d1-ddfe-46dc-aa7b-225708443379 [ 798.202107] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c629d946-9a9f-470d-b38d-fcbb108b9974 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.245193] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "refresh_cache-23cc444d-ff15-441b-8a0c-2ded49fd3cf0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.245603] env[61768]: DEBUG nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Instance network_info: |[{"id": "5405ba6b-c888-48a0-9af2-4024e9af563e", "address": "fa:16:3e:1c:2a:4f", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5405ba6b-c8", "ovs_interfaceid": "5405ba6b-c888-48a0-9af2-4024e9af563e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 798.246192] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1c:2a:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5405ba6b-c888-48a0-9af2-4024e9af563e', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 798.253896] env[61768]: DEBUG oslo.service.loopingcall [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.254811] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 798.255065] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8ded1e8c-e62e-43db-9317-ed6154ca0ecb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.280330] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 798.280330] env[61768]: value = "task-1228873" [ 798.280330] env[61768]: _type = "Task" [ 798.280330] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.289726] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228873, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.326081] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228872, 'name': Rename_Task, 'duration_secs': 0.200913} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.328249] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 798.329598] env[61768]: DEBUG nova.compute.manager [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Received event network-changed-5405ba6b-c888-48a0-9af2-4024e9af563e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 798.329800] env[61768]: DEBUG nova.compute.manager [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Refreshing instance network info cache due to event network-changed-5405ba6b-c888-48a0-9af2-4024e9af563e. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 798.330043] env[61768]: DEBUG oslo_concurrency.lockutils [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] Acquiring lock "refresh_cache-23cc444d-ff15-441b-8a0c-2ded49fd3cf0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.330186] env[61768]: DEBUG oslo_concurrency.lockutils [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] Acquired lock "refresh_cache-23cc444d-ff15-441b-8a0c-2ded49fd3cf0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.330363] env[61768]: DEBUG nova.network.neutron [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Refreshing network info cache for port 5405ba6b-c888-48a0-9af2-4024e9af563e {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 798.331528] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d9386edb-f7f6-4f9e-bfce-28a05d59bc0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.341205] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 798.341205] env[61768]: value = "task-1228874" [ 798.341205] env[61768]: _type = "Task" [ 798.341205] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.354886] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228874, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.384954] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d0e31-d6f0-f029-d0f8-e4eed1ffd567, 'name': SearchDatastore_Task, 'duration_secs': 0.011054} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.385977] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfa1cd94-0899-41e3-8cca-2b1e19753489 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.392704] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 798.392704] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52eb85f4-9725-c2f9-7af2-1d277857e8e9" [ 798.392704] env[61768]: _type = "Task" [ 798.392704] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.402676] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52eb85f4-9725-c2f9-7af2-1d277857e8e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.446332] env[61768]: DEBUG nova.compute.manager [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Stashing vm_state: active {{(pid=61768) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 798.704894] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d02b03db-471b-4940-aca3-723d5696540f tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "76c282d1-ddfe-46dc-aa7b-225708443379" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.132s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.793184] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228873, 'name': CreateVM_Task, 'duration_secs': 0.426228} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.793416] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 798.794276] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.794496] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.794860] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 798.795197] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78e60d15-b853-4856-a367-04d2c3971ea4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.801473] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 798.801473] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529dba50-2205-9ccf-1b1f-b08e5a063d01" [ 798.801473] env[61768]: _type = "Task" [ 798.801473] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.812539] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529dba50-2205-9ccf-1b1f-b08e5a063d01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.855017] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228874, 'name': PowerOnVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.910312] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52eb85f4-9725-c2f9-7af2-1d277857e8e9, 'name': SearchDatastore_Task, 'duration_secs': 0.014236} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.910539] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.910784] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] a36b542c-334f-46b1-9030-ee5c9f8c55c1/a36b542c-334f-46b1-9030-ee5c9f8c55c1.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 798.911094] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c8ef7cb-e5e4-4417-9a93-fde08ae6ec44 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.921037] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 798.921037] env[61768]: value = "task-1228875" [ 798.921037] env[61768]: _type = "Task" [ 798.921037] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.935029] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228875, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.971665] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.128834] env[61768]: DEBUG nova.compute.manager [req-55bb44f8-f70d-4bde-8cd2-55c3617488ac req-966778f8-1b11-4ccb-b4a8-3faa38866b0d service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Received event network-vif-plugged-2889bdd9-2069-495a-b80a-b20bbec5ef00 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.129148] env[61768]: DEBUG oslo_concurrency.lockutils [req-55bb44f8-f70d-4bde-8cd2-55c3617488ac req-966778f8-1b11-4ccb-b4a8-3faa38866b0d service nova] Acquiring lock "d928f14c-b94a-4abf-a053-51015f3bc6c5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.129432] env[61768]: DEBUG oslo_concurrency.lockutils [req-55bb44f8-f70d-4bde-8cd2-55c3617488ac req-966778f8-1b11-4ccb-b4a8-3faa38866b0d service nova] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.129628] env[61768]: DEBUG oslo_concurrency.lockutils [req-55bb44f8-f70d-4bde-8cd2-55c3617488ac req-966778f8-1b11-4ccb-b4a8-3faa38866b0d service nova] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.129808] env[61768]: DEBUG nova.compute.manager [req-55bb44f8-f70d-4bde-8cd2-55c3617488ac req-966778f8-1b11-4ccb-b4a8-3faa38866b0d service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] No waiting events found dispatching network-vif-plugged-2889bdd9-2069-495a-b80a-b20bbec5ef00 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.129986] env[61768]: WARNING nova.compute.manager [req-55bb44f8-f70d-4bde-8cd2-55c3617488ac req-966778f8-1b11-4ccb-b4a8-3faa38866b0d service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Received unexpected event network-vif-plugged-2889bdd9-2069-495a-b80a-b20bbec5ef00 for instance with vm_state building and task_state spawning. [ 799.178733] env[61768]: DEBUG nova.network.neutron [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Updated VIF entry in instance network info cache for port 5405ba6b-c888-48a0-9af2-4024e9af563e. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 799.179156] env[61768]: DEBUG nova.network.neutron [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Updating instance_info_cache with network_info: [{"id": "5405ba6b-c888-48a0-9af2-4024e9af563e", "address": "fa:16:3e:1c:2a:4f", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5405ba6b-c8", "ovs_interfaceid": "5405ba6b-c888-48a0-9af2-4024e9af563e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.320059] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529dba50-2205-9ccf-1b1f-b08e5a063d01, 'name': SearchDatastore_Task, 'duration_secs': 0.011492} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.320059] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.320059] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 799.320243] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.320477] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.320874] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 799.321615] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-849007d5-3b8b-4c53-b946-6d17e31d1f9f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.343102] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 799.343546] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 799.348292] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee16686e-c246-404e-8532-27f81d0b7403 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.359440] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 799.359440] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52df159b-f526-39d2-1a9f-8856431de75e" [ 799.359440] env[61768]: _type = "Task" [ 799.359440] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.359650] env[61768]: DEBUG oslo_vmware.api [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1228874, 'name': PowerOnVM_Task, 'duration_secs': 0.710123} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.359926] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 799.360244] env[61768]: INFO nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Took 8.70 seconds to spawn the instance on the hypervisor. [ 799.360507] env[61768]: DEBUG nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.366224] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3c7bf82-bf2e-4e1d-b750-5bfd836852e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.384497] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52df159b-f526-39d2-1a9f-8856431de75e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.440487] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228875, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.525741] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b23ced6d-f9fa-4c64-8f32-b2cc514a0289 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.534870] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12257cb0-4ef0-49b1-af27-494c99b80bea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.576631] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9eae1a0-99e2-490c-b0e4-31c2f97b7e01 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.586599] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-195c11ae-6e03-4903-b496-fc6443b8c4e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.606969] env[61768]: DEBUG nova.compute.provider_tree [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.610362] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Successfully updated port: 2889bdd9-2069-495a-b80a-b20bbec5ef00 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.681843] env[61768]: DEBUG oslo_concurrency.lockutils [req-92044e84-255a-487d-bf54-32805a669f51 req-7be1ed8c-99f0-40d8-b01d-67cc523787a8 service nova] Releasing lock "refresh_cache-23cc444d-ff15-441b-8a0c-2ded49fd3cf0" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.871691] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52df159b-f526-39d2-1a9f-8856431de75e, 'name': SearchDatastore_Task, 'duration_secs': 0.025541} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.872590] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14655134-8aef-4783-aa1d-bbdbf93caf5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.880650] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 799.880650] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e5f9ef-1638-cd2a-b634-e88b210e9081" [ 799.880650] env[61768]: _type = "Task" [ 799.880650] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.895466] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e5f9ef-1638-cd2a-b634-e88b210e9081, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.896882] env[61768]: INFO nova.compute.manager [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Took 30.63 seconds to build instance. [ 799.933155] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228875, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619888} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.933689] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] a36b542c-334f-46b1-9030-ee5c9f8c55c1/a36b542c-334f-46b1-9030-ee5c9f8c55c1.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 799.933689] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.933973] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81a5b54c-153f-42a9-a473-f5c4e41f8e94 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.944285] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 799.944285] env[61768]: value = "task-1228876" [ 799.944285] env[61768]: _type = "Task" [ 799.944285] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.954150] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228876, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.074420] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "41f6300d-7462-4e11-b32b-8c892e87bafc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.074691] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.113862] env[61768]: DEBUG nova.scheduler.client.report [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.117226] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "refresh_cache-d928f14c-b94a-4abf-a053-51015f3bc6c5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.117226] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "refresh_cache-d928f14c-b94a-4abf-a053-51015f3bc6c5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.117226] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.352239] env[61768]: DEBUG nova.compute.manager [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Received event network-changed-2889bdd9-2069-495a-b80a-b20bbec5ef00 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 800.352239] env[61768]: DEBUG nova.compute.manager [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Refreshing instance network info cache due to event network-changed-2889bdd9-2069-495a-b80a-b20bbec5ef00. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 800.352239] env[61768]: DEBUG oslo_concurrency.lockutils [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] Acquiring lock "refresh_cache-d928f14c-b94a-4abf-a053-51015f3bc6c5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.397030] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e5f9ef-1638-cd2a-b634-e88b210e9081, 'name': SearchDatastore_Task, 'duration_secs': 0.027409} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.397191] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.397543] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 23cc444d-ff15-441b-8a0c-2ded49fd3cf0/23cc444d-ff15-441b-8a0c-2ded49fd3cf0.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 800.399159] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-66800135-76ef-4405-9cf1-aa6b019e9dc4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.400348] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9bbba95c-ee2f-438d-a80a-7d10811a62cd tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.157s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.407815] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 800.407815] env[61768]: value = "task-1228877" [ 800.407815] env[61768]: _type = "Task" [ 800.407815] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.417841] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228877, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.452742] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228876, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.118508} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.453051] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 800.453818] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2bfa2a-8fc5-4a60-a94e-e48b75914235 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.476218] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] a36b542c-334f-46b1-9030-ee5c9f8c55c1/a36b542c-334f-46b1-9030-ee5c9f8c55c1.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 800.476566] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-71154332-2c41-4d9b-ba62-73f37a2c98ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.498204] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 800.498204] env[61768]: value = "task-1228878" [ 800.498204] env[61768]: _type = "Task" [ 800.498204] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.507021] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228878, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.622026] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.622026] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.623566] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.099s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.625490] env[61768]: INFO nova.compute.claims [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.666266] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.808738] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Updating instance_info_cache with network_info: [{"id": "2889bdd9-2069-495a-b80a-b20bbec5ef00", "address": "fa:16:3e:69:b1:74", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2889bdd9-20", "ovs_interfaceid": "2889bdd9-2069-495a-b80a-b20bbec5ef00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.905067] env[61768]: DEBUG nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 800.919922] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228877, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.011020] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228878, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.134409] env[61768]: DEBUG nova.compute.utils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.141657] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.141939] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 801.219979] env[61768]: DEBUG nova.policy [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5721b9005cf8496995615cfc21da01a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '851a60883ea34523abf3a0ce81e6d40f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.311763] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "refresh_cache-d928f14c-b94a-4abf-a053-51015f3bc6c5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.312144] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Instance network_info: |[{"id": "2889bdd9-2069-495a-b80a-b20bbec5ef00", "address": "fa:16:3e:69:b1:74", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2889bdd9-20", "ovs_interfaceid": "2889bdd9-2069-495a-b80a-b20bbec5ef00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 801.312484] env[61768]: DEBUG oslo_concurrency.lockutils [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] Acquired lock "refresh_cache-d928f14c-b94a-4abf-a053-51015f3bc6c5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.312787] env[61768]: DEBUG nova.network.neutron [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Refreshing network info cache for port 2889bdd9-2069-495a-b80a-b20bbec5ef00 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 801.313926] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:b1:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2889bdd9-2069-495a-b80a-b20bbec5ef00', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 801.321753] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Creating folder: Project (851a60883ea34523abf3a0ce81e6d40f). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.324861] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-744acedc-3724-4bcb-adf7-dd40e10ee5f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.338060] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Created folder: Project (851a60883ea34523abf3a0ce81e6d40f) in parent group-v265360. [ 801.338262] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Creating folder: Instances. Parent ref: group-v265451. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 801.338605] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0e8243e9-08ac-41e0-9f96-5a57c45a1981 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.349281] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Created folder: Instances in parent group-v265451. [ 801.349539] env[61768]: DEBUG oslo.service.loopingcall [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.349765] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 801.349952] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a5e5f4b-3d0f-4719-b63c-e70a59723e0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.373223] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 801.373223] env[61768]: value = "task-1228881" [ 801.373223] env[61768]: _type = "Task" [ 801.373223] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.385242] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228881, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.429169] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228877, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.880494} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.429541] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 23cc444d-ff15-441b-8a0c-2ded49fd3cf0/23cc444d-ff15-441b-8a0c-2ded49fd3cf0.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 801.429788] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 801.430167] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fce1dac2-103e-421f-abbe-2c97f9449ed0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.439943] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 801.439943] env[61768]: value = "task-1228882" [ 801.439943] env[61768]: _type = "Task" [ 801.439943] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.446509] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 801.454468] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228882, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.511184] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228878, 'name': ReconfigVM_Task, 'duration_secs': 0.798749} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.511551] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Reconfigured VM instance instance-00000044 to attach disk [datastore1] a36b542c-334f-46b1-9030-ee5c9f8c55c1/a36b542c-334f-46b1-9030-ee5c9f8c55c1.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 801.512322] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33251447-3fe9-4496-b97e-8270c0c770a5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.519184] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 801.519184] env[61768]: value = "task-1228883" [ 801.519184] env[61768]: _type = "Task" [ 801.519184] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.529395] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228883, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.640708] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.709202] env[61768]: DEBUG nova.network.neutron [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Updated VIF entry in instance network info cache for port 2889bdd9-2069-495a-b80a-b20bbec5ef00. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 801.711985] env[61768]: DEBUG nova.network.neutron [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Updating instance_info_cache with network_info: [{"id": "2889bdd9-2069-495a-b80a-b20bbec5ef00", "address": "fa:16:3e:69:b1:74", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2889bdd9-20", "ovs_interfaceid": "2889bdd9-2069-495a-b80a-b20bbec5ef00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.714443] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Successfully created port: 97d08d00-af43-4a90-b759-76b75806aa51 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.884726] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228881, 'name': CreateVM_Task, 'duration_secs': 0.479441} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.887149] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 801.888081] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.888211] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.888570] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 801.888836] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac44d12a-4d49-4eb2-94ba-a1417374acb0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.893900] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 801.893900] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52069b42-2e72-ebbf-2794-fbf8c01d4122" [ 801.893900] env[61768]: _type = "Task" [ 801.893900] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.905010] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52069b42-2e72-ebbf-2794-fbf8c01d4122, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.952755] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228882, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078623} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.953075] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 801.953830] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe1249f-81ef-45fc-99cb-d80a08dbf5e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.977813] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 23cc444d-ff15-441b-8a0c-2ded49fd3cf0/23cc444d-ff15-441b-8a0c-2ded49fd3cf0.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 801.980689] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da68cec8-2d8b-491f-a082-2406b5e6925b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.002993] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 802.002993] env[61768]: value = "task-1228884" [ 802.002993] env[61768]: _type = "Task" [ 802.002993] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.016087] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228884, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.026395] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e006f2a-345c-4567-859f-077c56bf52d0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.037115] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f92ca48-03d2-490e-87ef-dea244eee587 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.040367] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228883, 'name': Rename_Task, 'duration_secs': 0.247885} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.040650] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 802.041255] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5bb2fc73-cc90-4eaf-aff9-560d73b73346 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.072144] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-911ca4e0-8812-482b-b227-92d21a74aa31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.074900] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 802.074900] env[61768]: value = "task-1228885" [ 802.074900] env[61768]: _type = "Task" [ 802.074900] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.081700] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e23e903-61c6-4f20-bb34-e312a707d16c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.089161] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228885, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.100661] env[61768]: DEBUG nova.compute.provider_tree [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 802.213332] env[61768]: DEBUG oslo_concurrency.lockutils [req-549b470d-617a-4c87-bc9d-a146bb108e75 req-01b229dc-b220-48ed-a0fb-bc2392bf38af service nova] Releasing lock "refresh_cache-d928f14c-b94a-4abf-a053-51015f3bc6c5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.405876] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52069b42-2e72-ebbf-2794-fbf8c01d4122, 'name': SearchDatastore_Task, 'duration_secs': 0.029445} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.406203] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.406473] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 802.406728] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.406883] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.407294] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 802.407364] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-372defba-0a21-454a-b0af-244593113d58 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.421199] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 802.421405] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 802.423700] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d4954cb-28eb-49df-a9ed-8b82f392e568 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.427511] env[61768]: DEBUG nova.compute.manager [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Received event network-changed-51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 802.427733] env[61768]: DEBUG nova.compute.manager [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Refreshing instance network info cache due to event network-changed-51b2319f-ceb1-41db-8019-1d918e70080d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 802.428106] env[61768]: DEBUG oslo_concurrency.lockutils [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] Acquiring lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.428378] env[61768]: DEBUG oslo_concurrency.lockutils [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] Acquired lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.428456] env[61768]: DEBUG nova.network.neutron [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Refreshing network info cache for port 51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 802.434694] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 802.434694] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f650e8-efe2-814d-f1e6-c9861058d619" [ 802.434694] env[61768]: _type = "Task" [ 802.434694] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.444829] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f650e8-efe2-814d-f1e6-c9861058d619, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.514405] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228884, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.585851] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228885, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.604069] env[61768]: DEBUG nova.scheduler.client.report [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.651261] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.680664] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.681074] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.681373] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.681717] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.682027] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.682335] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.682754] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.683117] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.683468] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.683802] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.684162] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.685638] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77d06aa-0994-4e7a-8c56-3edccadebd20 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.697273] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74c7ea6-7402-43f9-85dd-d44b8ed62b52 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.947495] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f650e8-efe2-814d-f1e6-c9861058d619, 'name': SearchDatastore_Task, 'duration_secs': 0.02337} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.948931] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3bfe4aab-f474-4048-9629-1344238557e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.955945] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 802.955945] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d0d09a-9690-9fc6-cfed-f7fd66f775b5" [ 802.955945] env[61768]: _type = "Task" [ 802.955945] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.967124] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d0d09a-9690-9fc6-cfed-f7fd66f775b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.015952] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228884, 'name': ReconfigVM_Task, 'duration_secs': 0.909581} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.016294] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 23cc444d-ff15-441b-8a0c-2ded49fd3cf0/23cc444d-ff15-441b-8a0c-2ded49fd3cf0.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.016944] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2649cca8-0ca7-4a01-a60c-0c84e0d31652 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.024813] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 803.024813] env[61768]: value = "task-1228886" [ 803.024813] env[61768]: _type = "Task" [ 803.024813] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.036825] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228886, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.087409] env[61768]: DEBUG oslo_vmware.api [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228885, 'name': PowerOnVM_Task, 'duration_secs': 0.946433} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.089903] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 803.090146] env[61768]: INFO nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Took 9.92 seconds to spawn the instance on the hypervisor. [ 803.090345] env[61768]: DEBUG nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.091165] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00c09336-b5a7-47f6-adcf-3977e794bf26 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.108699] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.485s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.109286] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 803.111940] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.653s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.112239] env[61768]: DEBUG nova.objects.instance [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lazy-loading 'resources' on Instance uuid 23a4450d-162a-4b2b-a009-7023851315a1 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 803.296573] env[61768]: DEBUG nova.network.neutron [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updated VIF entry in instance network info cache for port 51b2319f-ceb1-41db-8019-1d918e70080d. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 803.297033] env[61768]: DEBUG nova.network.neutron [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating instance_info_cache with network_info: [{"id": "51b2319f-ceb1-41db-8019-1d918e70080d", "address": "fa:16:3e:7d:47:ab", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b2319f-ce", "ovs_interfaceid": "51b2319f-ceb1-41db-8019-1d918e70080d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.479024] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d0d09a-9690-9fc6-cfed-f7fd66f775b5, 'name': SearchDatastore_Task, 'duration_secs': 0.017339} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.479024] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.479024] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] d928f14c-b94a-4abf-a053-51015f3bc6c5/d928f14c-b94a-4abf-a053-51015f3bc6c5.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 803.479024] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d28972c-1e40-4297-9271-98fe578d2bf7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.491025] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 803.491025] env[61768]: value = "task-1228887" [ 803.491025] env[61768]: _type = "Task" [ 803.491025] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.501447] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228887, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.523571] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Successfully updated port: 97d08d00-af43-4a90-b759-76b75806aa51 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 803.535500] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228886, 'name': Rename_Task, 'duration_secs': 0.17373} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.535855] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 803.536205] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e665f69c-0e49-4922-a1ac-51729e0a004e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.543638] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 803.543638] env[61768]: value = "task-1228888" [ 803.543638] env[61768]: _type = "Task" [ 803.543638] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.554092] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228888, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.609490] env[61768]: INFO nova.compute.manager [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Took 32.90 seconds to build instance. [ 803.614835] env[61768]: DEBUG nova.compute.utils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.619328] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.619328] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 803.677237] env[61768]: DEBUG nova.policy [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5721b9005cf8496995615cfc21da01a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '851a60883ea34523abf3a0ce81e6d40f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.800112] env[61768]: DEBUG oslo_concurrency.lockutils [req-e270a9fd-69a4-46ec-b566-2302b14b9572 req-22118e50-1ddb-43c1-9649-b55df1a6d4c6 service nova] Releasing lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.994209] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Successfully created port: 9cef3fe6-e733-4c02-82f1-2d914170b440 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.008866] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228887, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.020491] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c4c54f-1169-4f64-afc2-79204c1c7656 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.029500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "refresh_cache-356a69b0-7898-440d-9473-7b4572ed2315" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.029734] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "refresh_cache-356a69b0-7898-440d-9473-7b4572ed2315" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.029905] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 804.039825] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8382eeae-12a9-478b-9bd2-4a20f43e766a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.090498] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69017830-811f-4960-9a1a-dad77971db37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.095144] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228888, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.105205] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f0f1975-3a5f-43c9-82f4-fb321b8b1623 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.111997] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bbb728cf-d9eb-4ee7-bbd7-1f300a195633 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.406s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.121910] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 804.124983] env[61768]: DEBUG nova.compute.provider_tree [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.478217] env[61768]: DEBUG nova.compute.manager [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Received event network-vif-plugged-97d08d00-af43-4a90-b759-76b75806aa51 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.478414] env[61768]: DEBUG oslo_concurrency.lockutils [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] Acquiring lock "356a69b0-7898-440d-9473-7b4572ed2315-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.478631] env[61768]: DEBUG oslo_concurrency.lockutils [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] Lock "356a69b0-7898-440d-9473-7b4572ed2315-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.478807] env[61768]: DEBUG oslo_concurrency.lockutils [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] Lock "356a69b0-7898-440d-9473-7b4572ed2315-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.479269] env[61768]: DEBUG nova.compute.manager [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] No waiting events found dispatching network-vif-plugged-97d08d00-af43-4a90-b759-76b75806aa51 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 804.479391] env[61768]: WARNING nova.compute.manager [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Received unexpected event network-vif-plugged-97d08d00-af43-4a90-b759-76b75806aa51 for instance with vm_state building and task_state spawning. [ 804.479570] env[61768]: DEBUG nova.compute.manager [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Received event network-changed-97d08d00-af43-4a90-b759-76b75806aa51 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.479732] env[61768]: DEBUG nova.compute.manager [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Refreshing instance network info cache due to event network-changed-97d08d00-af43-4a90-b759-76b75806aa51. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.479906] env[61768]: DEBUG oslo_concurrency.lockutils [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] Acquiring lock "refresh_cache-356a69b0-7898-440d-9473-7b4572ed2315" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.500147] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228887, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659943} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.500489] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] d928f14c-b94a-4abf-a053-51015f3bc6c5/d928f14c-b94a-4abf-a053-51015f3bc6c5.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 804.500726] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 804.501040] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e1ce020-9255-4c69-a882-2d18059fa99d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.509056] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 804.509056] env[61768]: value = "task-1228889" [ 804.509056] env[61768]: _type = "Task" [ 804.509056] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.518548] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228889, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.555464] env[61768]: DEBUG oslo_vmware.api [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228888, 'name': PowerOnVM_Task, 'duration_secs': 0.609293} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.555955] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 804.556059] env[61768]: INFO nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Took 8.90 seconds to spawn the instance on the hypervisor. [ 804.556279] env[61768]: DEBUG nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.557171] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df27c560-ccc3-4917-acd6-b4cff69bcd43 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.568312] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.664193] env[61768]: ERROR nova.scheduler.client.report [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] [req-e8b6356c-2260-47d8-ab86-10a6c59c3474] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e8b6356c-2260-47d8-ab86-10a6c59c3474"}]} [ 804.693078] env[61768]: DEBUG nova.scheduler.client.report [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 804.708009] env[61768]: DEBUG nova.scheduler.client.report [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 804.708217] env[61768]: DEBUG nova.compute.provider_tree [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.721842] env[61768]: DEBUG nova.scheduler.client.report [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 804.743325] env[61768]: DEBUG nova.scheduler.client.report [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 804.753064] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Updating instance_info_cache with network_info: [{"id": "97d08d00-af43-4a90-b759-76b75806aa51", "address": "fa:16:3e:30:8f:a8", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97d08d00-af", "ovs_interfaceid": "97d08d00-af43-4a90-b759-76b75806aa51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.008568] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bac11c-4871-43cd-b718-658f333ec6ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.023067] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228889, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074572} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.023925] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf754bd7-1572-4edc-b9dc-beb693caec0c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.027041] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 805.027766] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b45fbd5-6f1e-41b0-98c7-9f48795ee9bc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.051301] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] d928f14c-b94a-4abf-a053-51015f3bc6c5/d928f14c-b94a-4abf-a053-51015f3bc6c5.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 805.074480] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4139b855-63e7-4dcc-aff4-215d19ac8615 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.094056] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dde6677-34d2-40b7-a1e5-2f007ffdf3e7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.098642] env[61768]: INFO nova.compute.manager [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Took 29.93 seconds to build instance. [ 805.108269] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e680853-6656-4efb-9f41-9266923d2608 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.111533] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 805.111533] env[61768]: value = "task-1228890" [ 805.111533] env[61768]: _type = "Task" [ 805.111533] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.124092] env[61768]: DEBUG nova.compute.provider_tree [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.131663] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.135035] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 805.157826] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 805.158148] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 805.158328] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 805.158523] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 805.158677] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 805.158830] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 805.159113] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 805.159236] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 805.159413] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 805.159585] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 805.159764] env[61768]: DEBUG nova.virt.hardware [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 805.160646] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de5ed52-c57e-4b3b-b9c4-657294ec9229 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.169757] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338bd28d-245b-4b6a-a1f0-6a962257e7f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.259697] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "refresh_cache-356a69b0-7898-440d-9473-7b4572ed2315" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.260099] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Instance network_info: |[{"id": "97d08d00-af43-4a90-b759-76b75806aa51", "address": "fa:16:3e:30:8f:a8", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97d08d00-af", "ovs_interfaceid": "97d08d00-af43-4a90-b759-76b75806aa51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 805.261170] env[61768]: DEBUG nova.compute.manager [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.261513] env[61768]: DEBUG oslo_concurrency.lockutils [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] Acquired lock "refresh_cache-356a69b0-7898-440d-9473-7b4572ed2315" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.261691] env[61768]: DEBUG nova.network.neutron [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Refreshing network info cache for port 97d08d00-af43-4a90-b759-76b75806aa51 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 805.263031] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:8f:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97d08d00-af43-4a90-b759-76b75806aa51', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.273065] env[61768]: DEBUG oslo.service.loopingcall [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.273797] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddbe9303-0979-4130-87f6-a24eaf90207a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.277429] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 805.277682] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-507e283b-cdf0-4161-87b0-b3b9046d6494 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.303027] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.303027] env[61768]: value = "task-1228891" [ 805.303027] env[61768]: _type = "Task" [ 805.303027] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.314070] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228891, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.601145] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b2a71b07-b22e-4464-8224-7daf14c800f2 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.717s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.624130] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228890, 'name': ReconfigVM_Task, 'duration_secs': 0.441084} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.624558] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Reconfigured VM instance instance-00000046 to attach disk [datastore2] d928f14c-b94a-4abf-a053-51015f3bc6c5/d928f14c-b94a-4abf-a053-51015f3bc6c5.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 805.625360] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-947728f9-6f37-4c6a-a101-df36f0621c3e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.629193] env[61768]: DEBUG nova.scheduler.client.report [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.645145] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 805.645145] env[61768]: value = "task-1228892" [ 805.645145] env[61768]: _type = "Task" [ 805.645145] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.666087] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228892, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.778195] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Successfully updated port: 9cef3fe6-e733-4c02-82f1-2d914170b440 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.806239] env[61768]: INFO nova.compute.manager [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] instance snapshotting [ 805.812955] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57050913-fc9d-430a-af34-db6b9e957554 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.822977] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228891, 'name': CreateVM_Task, 'duration_secs': 0.456677} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.838022] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 805.839452] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.839641] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.839981] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 805.840910] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4032b8a0-f2e4-4c88-a2cf-daa9595ede8e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.846827] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e40ce1b4-6430-4da2-8e7a-8b297780d574 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.860074] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 805.860074] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fdeecd-3869-a60d-e2b4-a6993e1770de" [ 805.860074] env[61768]: _type = "Task" [ 805.860074] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.871251] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fdeecd-3869-a60d-e2b4-a6993e1770de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.064610] env[61768]: DEBUG nova.network.neutron [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Updated VIF entry in instance network info cache for port 97d08d00-af43-4a90-b759-76b75806aa51. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 806.065357] env[61768]: DEBUG nova.network.neutron [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Updating instance_info_cache with network_info: [{"id": "97d08d00-af43-4a90-b759-76b75806aa51", "address": "fa:16:3e:30:8f:a8", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97d08d00-af", "ovs_interfaceid": "97d08d00-af43-4a90-b759-76b75806aa51", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.139177] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.027s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.142512] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.778s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.144070] env[61768]: INFO nova.compute.claims [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 806.157580] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228892, 'name': Rename_Task, 'duration_secs': 0.198608} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.157931] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 806.158273] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a68f97ca-daea-4413-a647-56b0f0f5fa8a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.164722] env[61768]: INFO nova.scheduler.client.report [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Deleted allocations for instance 23a4450d-162a-4b2b-a009-7023851315a1 [ 806.167085] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 806.167085] env[61768]: value = "task-1228893" [ 806.167085] env[61768]: _type = "Task" [ 806.167085] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.180921] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.287928] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "refresh_cache-b6e4f743-5692-4b46-892c-6c8917ccef98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.288190] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "refresh_cache-b6e4f743-5692-4b46-892c-6c8917ccef98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.288395] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 806.358827] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 806.359221] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-80096979-3dcd-4673-a9a8-8b4c4ffadbc2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.373436] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fdeecd-3869-a60d-e2b4-a6993e1770de, 'name': SearchDatastore_Task, 'duration_secs': 0.016993} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.375120] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.375456] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.375731] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.375916] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.376141] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.376488] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 806.376488] env[61768]: value = "task-1228894" [ 806.376488] env[61768]: _type = "Task" [ 806.376488] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.376703] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-903436c4-1ee0-486c-98aa-2dd43e8bb590 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.387504] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228894, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.390149] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.390353] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 806.391095] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef0293ff-4801-454d-b9e0-f1102b89aa1c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.396766] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 806.396766] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b1a59a-4f79-2243-a7a1-99c86f846992" [ 806.396766] env[61768]: _type = "Task" [ 806.396766] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.406566] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b1a59a-4f79-2243-a7a1-99c86f846992, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.503196] env[61768]: DEBUG nova.compute.manager [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Received event network-vif-plugged-9cef3fe6-e733-4c02-82f1-2d914170b440 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.503451] env[61768]: DEBUG oslo_concurrency.lockutils [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] Acquiring lock "b6e4f743-5692-4b46-892c-6c8917ccef98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.503712] env[61768]: DEBUG oslo_concurrency.lockutils [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.503930] env[61768]: DEBUG oslo_concurrency.lockutils [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.504144] env[61768]: DEBUG nova.compute.manager [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] No waiting events found dispatching network-vif-plugged-9cef3fe6-e733-4c02-82f1-2d914170b440 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.504326] env[61768]: WARNING nova.compute.manager [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Received unexpected event network-vif-plugged-9cef3fe6-e733-4c02-82f1-2d914170b440 for instance with vm_state building and task_state spawning. [ 806.504540] env[61768]: DEBUG nova.compute.manager [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Received event network-changed-9cef3fe6-e733-4c02-82f1-2d914170b440 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.504750] env[61768]: DEBUG nova.compute.manager [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Refreshing instance network info cache due to event network-changed-9cef3fe6-e733-4c02-82f1-2d914170b440. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 806.504944] env[61768]: DEBUG oslo_concurrency.lockutils [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] Acquiring lock "refresh_cache-b6e4f743-5692-4b46-892c-6c8917ccef98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 806.567601] env[61768]: DEBUG oslo_concurrency.lockutils [req-e290d373-2c9d-4485-9710-c2c128b7371c req-345e2f70-b869-42d3-8544-75f22f1149ed service nova] Releasing lock "refresh_cache-356a69b0-7898-440d-9473-7b4572ed2315" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.678052] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f5a1a8f3-9f78-49e9-a01b-04917a3076a5 tempest-ServerRescueTestJSONUnderV235-1420514686 tempest-ServerRescueTestJSONUnderV235-1420514686-project-member] Lock "23a4450d-162a-4b2b-a009-7023851315a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.707s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.684781] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228893, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.819961] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "aebe41ad-496d-4a53-b023-c8df2bca04f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.820248] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.838418] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.890451] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228894, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.910047] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b1a59a-4f79-2243-a7a1-99c86f846992, 'name': SearchDatastore_Task, 'duration_secs': 0.016081} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.910996] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-754ae594-d4a5-4115-bdaf-ceda6204d669 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.918408] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 806.918408] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52680198-9299-37f1-01fd-f44eda36420c" [ 806.918408] env[61768]: _type = "Task" [ 806.918408] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.930675] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52680198-9299-37f1-01fd-f44eda36420c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.021354] env[61768]: DEBUG nova.network.neutron [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Updating instance_info_cache with network_info: [{"id": "9cef3fe6-e733-4c02-82f1-2d914170b440", "address": "fa:16:3e:4a:16:93", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cef3fe6-e7", "ovs_interfaceid": "9cef3fe6-e733-4c02-82f1-2d914170b440", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.039951] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52761ef9-9a27-a2d0-1c84-993bba1d4b81/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 807.041453] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e0ac368-3568-40ee-a90c-0d80724404bb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.049707] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52761ef9-9a27-a2d0-1c84-993bba1d4b81/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 807.049917] env[61768]: ERROR oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52761ef9-9a27-a2d0-1c84-993bba1d4b81/disk-0.vmdk due to incomplete transfer. [ 807.050186] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3e5002c9-dc0e-451b-bf7b-c110b67d77ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.058329] env[61768]: DEBUG oslo_vmware.rw_handles [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52761ef9-9a27-a2d0-1c84-993bba1d4b81/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 807.058540] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Uploaded image 6c2ff2ef-4663-4449-a0eb-4863e7b27db2 to the Glance image server {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 807.060712] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 807.060988] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-7705f65c-273d-4993-a31b-f0ca35dcc679 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.068478] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 807.068478] env[61768]: value = "task-1228895" [ 807.068478] env[61768]: _type = "Task" [ 807.068478] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.078419] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228895, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.180629] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228893, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.323199] env[61768]: DEBUG nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 807.392845] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228894, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.431243] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52680198-9299-37f1-01fd-f44eda36420c, 'name': SearchDatastore_Task, 'duration_secs': 0.014461} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.431243] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.431349] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 356a69b0-7898-440d-9473-7b4572ed2315/356a69b0-7898-440d-9473-7b4572ed2315.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 807.431599] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3b49ad97-831a-4a48-bbbb-54ac5bd73a90 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.438926] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 807.438926] env[61768]: value = "task-1228896" [ 807.438926] env[61768]: _type = "Task" [ 807.438926] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.452243] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228896, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.523958] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "refresh_cache-b6e4f743-5692-4b46-892c-6c8917ccef98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.524374] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Instance network_info: |[{"id": "9cef3fe6-e733-4c02-82f1-2d914170b440", "address": "fa:16:3e:4a:16:93", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cef3fe6-e7", "ovs_interfaceid": "9cef3fe6-e733-4c02-82f1-2d914170b440", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.525022] env[61768]: DEBUG oslo_concurrency.lockutils [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] Acquired lock "refresh_cache-b6e4f743-5692-4b46-892c-6c8917ccef98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.525022] env[61768]: DEBUG nova.network.neutron [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Refreshing network info cache for port 9cef3fe6-e733-4c02-82f1-2d914170b440 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 807.530257] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:16:93', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0636c3f6-fcb7-4954-ab07-c5cd0dee37b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9cef3fe6-e733-4c02-82f1-2d914170b440', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.536372] env[61768]: DEBUG oslo.service.loopingcall [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.538693] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 807.539834] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805ba044-e11a-4808-a97e-2ed213611e91 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.542940] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b1ba8a55-4894-463b-99e9-9ae813a90bc9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.564692] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de185cf5-bbb7-469f-a9ba-e642fa053e88 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.567880] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.567880] env[61768]: value = "task-1228897" [ 807.567880] env[61768]: _type = "Task" [ 807.567880] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.601489] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-185cadc6-f4d4-4735-93c6-5b30ac24d4b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.607576] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228897, 'name': CreateVM_Task} progress is 15%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.612096] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228895, 'name': Destroy_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.616547] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b840c648-5ba6-4499-a46e-74f8fd6112c6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.632944] env[61768]: DEBUG nova.compute.provider_tree [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 807.681388] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228893, 'name': PowerOnVM_Task, 'duration_secs': 1.19487} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.681544] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 807.681698] env[61768]: INFO nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Took 9.52 seconds to spawn the instance on the hypervisor. [ 807.681944] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 807.682827] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd09b50d-3eb1-48c1-bf0b-5dcebf10919f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.849660] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 807.895129] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228894, 'name': CreateSnapshot_Task, 'duration_secs': 1.065454} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.895129] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 807.896303] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0e9d41-4fec-48e5-baca-d95c25602c23 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.950721] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228896, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505631} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.951878] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 356a69b0-7898-440d-9473-7b4572ed2315/356a69b0-7898-440d-9473-7b4572ed2315.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 807.952048] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 807.952403] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c237703-8bc9-466b-99f0-0008a3f6a453 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.961053] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 807.961053] env[61768]: value = "task-1228898" [ 807.961053] env[61768]: _type = "Task" [ 807.961053] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.970480] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228898, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.082472] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228895, 'name': Destroy_Task, 'duration_secs': 0.643469} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.085980] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Destroyed the VM [ 808.086129] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 808.086324] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228897, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.086538] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5c769b2c-aa6a-467e-89b0-a8d188769cd8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.095243] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 808.095243] env[61768]: value = "task-1228899" [ 808.095243] env[61768]: _type = "Task" [ 808.095243] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.103984] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228899, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.136793] env[61768]: DEBUG nova.scheduler.client.report [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 808.205726] env[61768]: INFO nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Took 26.34 seconds to build instance. [ 808.370521] env[61768]: DEBUG nova.network.neutron [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Updated VIF entry in instance network info cache for port 9cef3fe6-e733-4c02-82f1-2d914170b440. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 808.371132] env[61768]: DEBUG nova.network.neutron [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Updating instance_info_cache with network_info: [{"id": "9cef3fe6-e733-4c02-82f1-2d914170b440", "address": "fa:16:3e:4a:16:93", "network": {"id": "52140648-2f04-4e98-84a5-c0d3e2d43ffc", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-973552600-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "851a60883ea34523abf3a0ce81e6d40f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0636c3f6-fcb7-4954-ab07-c5cd0dee37b0", "external-id": "nsx-vlan-transportzone-857", "segmentation_id": 857, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9cef3fe6-e7", "ovs_interfaceid": "9cef3fe6-e733-4c02-82f1-2d914170b440", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.419522] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 808.420806] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0845ae51-579f-4c8f-869d-d160a95022ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.434084] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 808.434084] env[61768]: value = "task-1228900" [ 808.434084] env[61768]: _type = "Task" [ 808.434084] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.443851] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228900, 'name': CloneVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.470843] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228898, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.585621] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228897, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.608063] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228899, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.643996] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.644545] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 808.648151] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.722s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.648418] env[61768]: DEBUG nova.objects.instance [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lazy-loading 'resources' on Instance uuid a428f5e0-ade1-4aa6-af9d-0e33efcfec62 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 808.711116] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.588s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.874584] env[61768]: DEBUG oslo_concurrency.lockutils [req-4780de36-f271-4922-9b9e-77e0e999b77a req-4670dbbb-45de-4a3d-9d61-c45f4858ffa9 service nova] Releasing lock "refresh_cache-b6e4f743-5692-4b46-892c-6c8917ccef98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.944554] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228900, 'name': CloneVM_Task} progress is 93%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.972835] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228898, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.084569] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228897, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.109198] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228899, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.152440] env[61768]: DEBUG nova.compute.utils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 809.163246] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 809.163246] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 809.246162] env[61768]: DEBUG nova.policy [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1600e5a4cc4e44cb8f6d384d6fa1028a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54b7f349107b4f56960cae84edbea9e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 809.256707] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "86e24eb9-6a55-4658-9e62-64713b0289d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.256981] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.446632] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228900, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.473670] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228898, 'name': ExtendVirtualDisk_Task, 'duration_secs': 1.097793} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.474275] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.476772] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-000cd2b6-c3ba-4a01-ba37-2b03f0679f0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.504158] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 356a69b0-7898-440d-9473-7b4572ed2315/356a69b0-7898-440d-9473-7b4572ed2315.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.506478] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9217e274-491a-482d-81b7-1c986c8680a7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.531501] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Successfully created port: 667907df-2454-4a31-9b6f-4e202b308713 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 809.535818] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 809.535818] env[61768]: value = "task-1228901" [ 809.535818] env[61768]: _type = "Task" [ 809.535818] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.548156] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228901, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.576346] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d70e9e-4197-44e9-a302-c0ad6158e8aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.588739] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b401a28d-5c9a-4d4c-8344-94eb36a9a200 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.592372] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228897, 'name': CreateVM_Task, 'duration_secs': 1.658441} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.592561] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 809.594224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.594224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.594429] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 809.594554] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3854eb3-dae6-4147-bc6c-aca06f444c0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.625313] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba4d44fe-30bd-48c8-ad08-e8670b2484e2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.629133] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 809.629133] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c90461-b271-37eb-ca44-902dc4e3def6" [ 809.629133] env[61768]: _type = "Task" [ 809.629133] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.637870] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228899, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.640314] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36f288f-bc4b-433a-ac53-30e17b4cae80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.647018] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c90461-b271-37eb-ca44-902dc4e3def6, 'name': SearchDatastore_Task, 'duration_secs': 0.010206} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.647726] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.648014] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.648307] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.648560] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.649026] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.649026] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cd91616-9184-4680-87c8-ad597aaf2d8c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.660347] env[61768]: DEBUG nova.compute.provider_tree [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 809.662712] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 809.669951] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.672024] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 809.672024] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e51458b-5f97-4927-a512-82d9251618d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.678632] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 809.678632] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5224ae90-c9e6-e4d1-2530-a2e613a3bf2a" [ 809.678632] env[61768]: _type = "Task" [ 809.678632] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.688458] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5224ae90-c9e6-e4d1-2530-a2e613a3bf2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.759213] env[61768]: DEBUG nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.946423] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228900, 'name': CloneVM_Task} progress is 95%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.046577] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228901, 'name': ReconfigVM_Task, 'duration_secs': 0.341031} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.046577] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 356a69b0-7898-440d-9473-7b4572ed2315/356a69b0-7898-440d-9473-7b4572ed2315.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.047303] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-95ba001c-1396-4642-8002-5b3ac71b6d0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.057714] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 810.057714] env[61768]: value = "task-1228902" [ 810.057714] env[61768]: _type = "Task" [ 810.057714] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.068013] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228902, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.131553] env[61768]: DEBUG oslo_vmware.api [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1228899, 'name': RemoveSnapshot_Task, 'duration_secs': 1.65494} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.131878] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 810.132547] env[61768]: INFO nova.compute.manager [None req-17223784-ed28-4587-9049-b37d4d823f9f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Took 17.53 seconds to snapshot the instance on the hypervisor. [ 810.190418] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5224ae90-c9e6-e4d1-2530-a2e613a3bf2a, 'name': SearchDatastore_Task, 'duration_secs': 0.010796} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.191786] env[61768]: ERROR nova.scheduler.client.report [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] [req-db348f27-0302-457f-ace8-0f0d0e5d208f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-db348f27-0302-457f-ace8-0f0d0e5d208f"}]} [ 810.195024] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f8cf847-87ee-47fd-9d20-e152de709876 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.202885] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 810.202885] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f1af3d-7f10-a349-76bf-8cf7e7d849c4" [ 810.202885] env[61768]: _type = "Task" [ 810.202885] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.209875] env[61768]: DEBUG nova.scheduler.client.report [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 810.215546] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f1af3d-7f10-a349-76bf-8cf7e7d849c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.230408] env[61768]: DEBUG nova.scheduler.client.report [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 810.230762] env[61768]: DEBUG nova.compute.provider_tree [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.243034] env[61768]: DEBUG nova.scheduler.client.report [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 810.261445] env[61768]: DEBUG nova.scheduler.client.report [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 810.281008] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.448432] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228900, 'name': CloneVM_Task, 'duration_secs': 1.642488} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.448755] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Created linked-clone VM from snapshot [ 810.449531] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0e30e8-dc29-4051-8697-aa05b28ff42e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.459045] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Uploading image b3123ef7-ca2d-484f-8ef6-476cbadf1385 {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 810.472125] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 810.472746] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-386b4ba0-7f75-4115-aedd-b0ff3d2c695a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.480987] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 810.480987] env[61768]: value = "task-1228903" [ 810.480987] env[61768]: _type = "Task" [ 810.480987] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.489806] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228903, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.552495] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38664599-6da3-4090-8013-3ec737faad8a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.564493] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94d449fb-1229-4525-8f63-f308d8ee6980 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.572975] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228902, 'name': Rename_Task, 'duration_secs': 0.149046} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.596247] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 810.596787] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b9c97842-e9a6-4d79-8065-8c42ec1c7426 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.599096] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62451dfa-538c-4667-97e2-6bc4938f9499 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.609882] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722717cf-4096-49a4-8947-4fbaa6fefedc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.613719] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 810.613719] env[61768]: value = "task-1228904" [ 810.613719] env[61768]: _type = "Task" [ 810.613719] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.624972] env[61768]: DEBUG nova.compute.provider_tree [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.631905] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228904, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.674794] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 810.699154] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 810.699451] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 810.699587] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 810.699778] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 810.699945] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 810.700201] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 810.700333] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 810.700528] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 810.700723] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 810.701058] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 810.701103] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 810.701946] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fcd9a6-09e5-4403-a146-dd70b4388277 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.716759] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93fe7c75-bacd-41fc-8c59-8d58ed65f9a7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.720498] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f1af3d-7f10-a349-76bf-8cf7e7d849c4, 'name': SearchDatastore_Task, 'duration_secs': 0.021231} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.720762] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.721028] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] b6e4f743-5692-4b46-892c-6c8917ccef98/b6e4f743-5692-4b46-892c-6c8917ccef98.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 810.721949] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d1976831-13e1-4f0c-aeb4-ca5d4f7ac17a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.738337] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 810.738337] env[61768]: value = "task-1228905" [ 810.738337] env[61768]: _type = "Task" [ 810.738337] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.749566] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228905, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.996384] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228903, 'name': Destroy_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.035947] env[61768]: DEBUG nova.compute.manager [req-8fc0431e-3222-49f0-bf30-5e780d2b56f5 req-5bbf9e2b-91b2-45c9-90dc-c220ec996758 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Received event network-vif-plugged-667907df-2454-4a31-9b6f-4e202b308713 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 811.036448] env[61768]: DEBUG oslo_concurrency.lockutils [req-8fc0431e-3222-49f0-bf30-5e780d2b56f5 req-5bbf9e2b-91b2-45c9-90dc-c220ec996758 service nova] Acquiring lock "d4a765bd-a56d-44b7-8db3-c081832d58c9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 811.036728] env[61768]: DEBUG oslo_concurrency.lockutils [req-8fc0431e-3222-49f0-bf30-5e780d2b56f5 req-5bbf9e2b-91b2-45c9-90dc-c220ec996758 service nova] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.036970] env[61768]: DEBUG oslo_concurrency.lockutils [req-8fc0431e-3222-49f0-bf30-5e780d2b56f5 req-5bbf9e2b-91b2-45c9-90dc-c220ec996758 service nova] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.037214] env[61768]: DEBUG nova.compute.manager [req-8fc0431e-3222-49f0-bf30-5e780d2b56f5 req-5bbf9e2b-91b2-45c9-90dc-c220ec996758 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] No waiting events found dispatching network-vif-plugged-667907df-2454-4a31-9b6f-4e202b308713 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 811.037455] env[61768]: WARNING nova.compute.manager [req-8fc0431e-3222-49f0-bf30-5e780d2b56f5 req-5bbf9e2b-91b2-45c9-90dc-c220ec996758 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Received unexpected event network-vif-plugged-667907df-2454-4a31-9b6f-4e202b308713 for instance with vm_state building and task_state spawning. [ 811.127758] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228904, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.171030] env[61768]: DEBUG nova.scheduler.client.report [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Updated inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with generation 89 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 811.171030] env[61768]: DEBUG nova.compute.provider_tree [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 89 to 90 during operation: update_inventory {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 811.171030] env[61768]: DEBUG nova.compute.provider_tree [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.173507] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Successfully updated port: 667907df-2454-4a31-9b6f-4e202b308713 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.250461] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228905, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.493984] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228903, 'name': Destroy_Task, 'duration_secs': 0.665436} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.494129] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Destroyed the VM [ 811.494332] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 811.494597] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8abba3db-e927-44d9-86f0-0818eb1534df {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.501914] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 811.501914] env[61768]: value = "task-1228906" [ 811.501914] env[61768]: _type = "Task" [ 811.501914] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.510506] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228906, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.624537] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228904, 'name': PowerOnVM_Task, 'duration_secs': 0.576322} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.624802] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 811.625036] env[61768]: INFO nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Took 8.97 seconds to spawn the instance on the hypervisor. [ 811.625216] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.626088] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072698c9-d3c9-46ac-bfc4-01549e777230 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.675856] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.028s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.678176] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.938s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.678434] env[61768]: DEBUG nova.objects.instance [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lazy-loading 'resources' on Instance uuid 77319462-e447-405b-9269-82581effe005 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.679728] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "refresh_cache-d4a765bd-a56d-44b7-8db3-c081832d58c9" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.679864] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "refresh_cache-d4a765bd-a56d-44b7-8db3-c081832d58c9" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.680028] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.698042] env[61768]: INFO nova.scheduler.client.report [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Deleted allocations for instance a428f5e0-ade1-4aa6-af9d-0e33efcfec62 [ 811.750377] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228905, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.539574} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.750644] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] b6e4f743-5692-4b46-892c-6c8917ccef98/b6e4f743-5692-4b46-892c-6c8917ccef98.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 811.750876] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 811.751149] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a5298181-4183-41d9-8961-2bc862e54aeb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.757934] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 811.757934] env[61768]: value = "task-1228907" [ 811.757934] env[61768]: _type = "Task" [ 811.757934] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.766991] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228907, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.012562] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228906, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.141945] env[61768]: INFO nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Took 29.23 seconds to build instance. [ 812.208640] env[61768]: DEBUG oslo_concurrency.lockutils [None req-534afcb6-8e7f-4a5c-b853-cd28815587bb tempest-ServersNegativeTestMultiTenantJSON-1994332773 tempest-ServersNegativeTestMultiTenantJSON-1994332773-project-member] Lock "a428f5e0-ade1-4aa6-af9d-0e33efcfec62" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.347s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.215798] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.272958] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228907, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.183544} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.272958] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 812.273702] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60128e9b-a16d-4de2-ba8b-b716c44c5528 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.300952] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Reconfiguring VM instance instance-00000048 to attach disk [datastore2] b6e4f743-5692-4b46-892c-6c8917ccef98/b6e4f743-5692-4b46-892c-6c8917ccef98.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 812.303907] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff706951-e91c-4004-87b2-0ced9a78c40f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.327259] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 812.327259] env[61768]: value = "task-1228908" [ 812.327259] env[61768]: _type = "Task" [ 812.327259] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.338641] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228908, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.455125] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Updating instance_info_cache with network_info: [{"id": "667907df-2454-4a31-9b6f-4e202b308713", "address": "fa:16:3e:4b:44:b1", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap667907df-24", "ovs_interfaceid": "667907df-2454-4a31-9b6f-4e202b308713", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.515125] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228906, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.541332] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b9f31a-9718-434f-8d5f-7d9f8dac19ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.549658] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be01e939-3735-4856-9a7d-a48b89656920 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.581851] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc85814-5cdd-472a-8c80-70da7f36c0af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.590172] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cad7dbaf-35e8-4f3d-9a47-408c189c9dc9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.604515] env[61768]: DEBUG nova.compute.provider_tree [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.644611] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "356a69b0-7898-440d-9473-7b4572ed2315" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.491s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.841938] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228908, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.960032] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "refresh_cache-d4a765bd-a56d-44b7-8db3-c081832d58c9" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.960032] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Instance network_info: |[{"id": "667907df-2454-4a31-9b6f-4e202b308713", "address": "fa:16:3e:4b:44:b1", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap667907df-24", "ovs_interfaceid": "667907df-2454-4a31-9b6f-4e202b308713", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 812.960032] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:44:b1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31a7f15-a808-4199-9071-31fd05e316ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '667907df-2454-4a31-9b6f-4e202b308713', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 812.973441] env[61768]: DEBUG oslo.service.loopingcall [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 812.974804] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 812.974804] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-173bdb8d-eb57-4df8-9a49-98026b6704c9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.000883] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.000883] env[61768]: value = "task-1228909" [ 813.000883] env[61768]: _type = "Task" [ 813.000883] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.018419] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228909, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.023892] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228906, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.075079] env[61768]: DEBUG nova.compute.manager [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Received event network-changed-667907df-2454-4a31-9b6f-4e202b308713 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 813.075412] env[61768]: DEBUG nova.compute.manager [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Refreshing instance network info cache due to event network-changed-667907df-2454-4a31-9b6f-4e202b308713. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 813.077952] env[61768]: DEBUG oslo_concurrency.lockutils [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] Acquiring lock "refresh_cache-d4a765bd-a56d-44b7-8db3-c081832d58c9" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.077952] env[61768]: DEBUG oslo_concurrency.lockutils [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] Acquired lock "refresh_cache-d4a765bd-a56d-44b7-8db3-c081832d58c9" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.077952] env[61768]: DEBUG nova.network.neutron [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Refreshing network info cache for port 667907df-2454-4a31-9b6f-4e202b308713 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 813.107235] env[61768]: DEBUG nova.scheduler.client.report [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 813.339580] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228908, 'name': ReconfigVM_Task, 'duration_secs': 0.83424} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.339950] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Reconfigured VM instance instance-00000048 to attach disk [datastore2] b6e4f743-5692-4b46-892c-6c8917ccef98/b6e4f743-5692-4b46-892c-6c8917ccef98.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 813.340543] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ea44071-6096-43d6-8ec0-f36115b532b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.348644] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 813.348644] env[61768]: value = "task-1228910" [ 813.348644] env[61768]: _type = "Task" [ 813.348644] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.358031] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228910, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.513069] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228909, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.518606] env[61768]: DEBUG oslo_vmware.api [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228906, 'name': RemoveSnapshot_Task, 'duration_secs': 1.705098} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.518892] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 813.612659] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.934s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.615672] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.563s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.617251] env[61768]: INFO nova.compute.claims [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.644140] env[61768]: INFO nova.scheduler.client.report [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Deleted allocations for instance 77319462-e447-405b-9269-82581effe005 [ 813.859980] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228910, 'name': Rename_Task, 'duration_secs': 0.322531} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.860367] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 813.860679] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8741c543-6d25-4f99-84a7-dbc8ba0e0f77 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.871150] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 813.871150] env[61768]: value = "task-1228911" [ 813.871150] env[61768]: _type = "Task" [ 813.871150] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.879705] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.014083] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228909, 'name': CreateVM_Task, 'duration_secs': 0.618874} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.014599] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.016167] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.017220] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.017733] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.018909] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e77069c-a5d8-4fb7-b5e0-6527fa235a5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.023923] env[61768]: WARNING nova.compute.manager [None req-a63cf0fa-acfa-42a8-8e8c-99160e120c6c tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Image not found during snapshot: nova.exception.ImageNotFound: Image b3123ef7-ca2d-484f-8ef6-476cbadf1385 could not be found. [ 814.031060] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 814.031060] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522426b9-ca89-e480-6f0c-a5edef2b7385" [ 814.031060] env[61768]: _type = "Task" [ 814.031060] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.041667] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522426b9-ca89-e480-6f0c-a5edef2b7385, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.153915] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442648c8-b27c-4adb-a5c2-6dc161ffd126 tempest-ServersTestMultiNic-1984043383 tempest-ServersTestMultiNic-1984043383-project-member] Lock "77319462-e447-405b-9269-82581effe005" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.928s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.173160] env[61768]: DEBUG nova.network.neutron [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Updated VIF entry in instance network info cache for port 667907df-2454-4a31-9b6f-4e202b308713. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 814.173544] env[61768]: DEBUG nova.network.neutron [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Updating instance_info_cache with network_info: [{"id": "667907df-2454-4a31-9b6f-4e202b308713", "address": "fa:16:3e:4b:44:b1", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap667907df-24", "ovs_interfaceid": "667907df-2454-4a31-9b6f-4e202b308713", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.382651] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228911, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.456752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.456752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.456752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.456752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.456752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.459039] env[61768]: INFO nova.compute.manager [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Terminating instance [ 814.460486] env[61768]: DEBUG nova.compute.manager [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 814.460727] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 814.461594] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f92f01-00fc-463c-9ba3-067d9c0c658d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.470134] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 814.470290] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bfc8c22-f7d3-440e-bb3a-2f8ade3e386e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.478933] env[61768]: DEBUG oslo_vmware.api [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 814.478933] env[61768]: value = "task-1228912" [ 814.478933] env[61768]: _type = "Task" [ 814.478933] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.487412] env[61768]: DEBUG oslo_vmware.api [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.547663] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522426b9-ca89-e480-6f0c-a5edef2b7385, 'name': SearchDatastore_Task, 'duration_secs': 0.01218} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.548138] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.548424] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.548681] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.548843] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.549028] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.549342] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-88687766-4932-4965-86a8-1f16af9c65f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.559458] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.559655] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 814.560447] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f974405-0577-406c-935e-37ab03eb7ea0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.566144] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 814.566144] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52781e20-86c4-b43f-9a8d-8dec30b507d9" [ 814.566144] env[61768]: _type = "Task" [ 814.566144] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.574956] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52781e20-86c4-b43f-9a8d-8dec30b507d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.679274] env[61768]: DEBUG oslo_concurrency.lockutils [req-265d815d-54c0-4fa3-a65d-44545b1f01c7 req-7d6ebb15-ea3a-4b77-bc0d-e1b7bb752468 service nova] Releasing lock "refresh_cache-d4a765bd-a56d-44b7-8db3-c081832d58c9" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.881765] env[61768]: DEBUG oslo_vmware.api [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228911, 'name': PowerOnVM_Task, 'duration_secs': 0.848117} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.882073] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 814.882161] env[61768]: INFO nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Took 9.75 seconds to spawn the instance on the hypervisor. [ 814.882340] env[61768]: DEBUG nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.883253] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41be3c99-b902-4168-9fac-9ad90e15ec73 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.956120] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce34e9c3-835f-46c7-897d-75d6dadfc6f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.964365] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207156f1-9a84-4a86-8415-82cc983e8564 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.002768] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b855ad-efb4-46ac-abab-5caef56441af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.011325] env[61768]: DEBUG oslo_vmware.api [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228912, 'name': PowerOffVM_Task, 'duration_secs': 0.303309} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.013599] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 815.013808] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 815.014132] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4640822b-8150-4ee7-ba54-50dc51679628 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.016859] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f52e6aa-402b-4cf8-9b60-3a52386c588f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.032472] env[61768]: DEBUG nova.compute.provider_tree [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.078424] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52781e20-86c4-b43f-9a8d-8dec30b507d9, 'name': SearchDatastore_Task, 'duration_secs': 0.013548} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.079248] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c2fdd75-1a5b-499c-87b3-8c62423eeedd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.085908] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 815.085908] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522e52f2-30f1-f149-760b-991446f484f6" [ 815.085908] env[61768]: _type = "Task" [ 815.085908] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.097937] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522e52f2-30f1-f149-760b-991446f484f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.099569] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 815.099569] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 815.100637] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleting the datastore file [datastore1] a36b542c-334f-46b1-9030-ee5c9f8c55c1 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.100637] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fcefc43e-9868-4f33-ab43-56facca231e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.107677] env[61768]: DEBUG oslo_vmware.api [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 815.107677] env[61768]: value = "task-1228914" [ 815.107677] env[61768]: _type = "Task" [ 815.107677] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.117830] env[61768]: DEBUG oslo_vmware.api [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.416266] env[61768]: INFO nova.compute.manager [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Took 30.92 seconds to build instance. [ 815.538018] env[61768]: DEBUG nova.scheduler.client.report [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.601259] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522e52f2-30f1-f149-760b-991446f484f6, 'name': SearchDatastore_Task, 'duration_secs': 0.018827} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.601903] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.602499] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] d4a765bd-a56d-44b7-8db3-c081832d58c9/d4a765bd-a56d-44b7-8db3-c081832d58c9.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 815.603618] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0969504d-1043-4078-ac27-cf8871c57915 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.623836] env[61768]: DEBUG oslo_vmware.api [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1228914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.310275} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.623836] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 815.623836] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 815.623836] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 815.623836] env[61768]: INFO nova.compute.manager [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Took 1.16 seconds to destroy the instance on the hypervisor. [ 815.623836] env[61768]: DEBUG oslo.service.loopingcall [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.628300] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 815.628300] env[61768]: value = "task-1228915" [ 815.628300] env[61768]: _type = "Task" [ 815.628300] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.629460] env[61768]: DEBUG nova.compute.manager [-] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.629460] env[61768]: DEBUG nova.network.neutron [-] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 815.642592] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228915, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.918145] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d01e30c2-b28b-4136-81e4-02ecd4c13958 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.738s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.044477] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.045207] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.050797] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.027s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.050797] env[61768]: DEBUG nova.objects.instance [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lazy-loading 'resources' on Instance uuid fbb8c9db-e823-444a-84de-b231cf9b8a8a {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.145284] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228915, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.181304] env[61768]: DEBUG nova.compute.manager [req-0543b15c-0251-4a0c-8f7a-5b1df1bf6178 req-4c2ab7ce-3b36-49a5-bc74-7b724f158bf9 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Received event network-vif-deleted-c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 816.181304] env[61768]: INFO nova.compute.manager [req-0543b15c-0251-4a0c-8f7a-5b1df1bf6178 req-4c2ab7ce-3b36-49a5-bc74-7b724f158bf9 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Neutron deleted interface c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10; detaching it from the instance and deleting it from the info cache [ 816.181304] env[61768]: DEBUG nova.network.neutron [req-0543b15c-0251-4a0c-8f7a-5b1df1bf6178 req-4c2ab7ce-3b36-49a5-bc74-7b724f158bf9 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.553771] env[61768]: DEBUG nova.compute.utils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 816.558104] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 816.558436] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.576736] env[61768]: DEBUG nova.network.neutron [-] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.612460] env[61768]: DEBUG nova.policy [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1600e5a4cc4e44cb8f6d384d6fa1028a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '54b7f349107b4f56960cae84edbea9e0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.646470] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228915, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.856704} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.646470] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] d4a765bd-a56d-44b7-8db3-c081832d58c9/d4a765bd-a56d-44b7-8db3-c081832d58c9.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 816.646470] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.646470] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-aa958f7e-43e3-4cca-8e9d-c9cfb68dc3d5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.654166] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 816.654166] env[61768]: value = "task-1228916" [ 816.654166] env[61768]: _type = "Task" [ 816.654166] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.663794] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228916, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.684881] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba08572a-a7f2-494c-a9ad-45878ad562ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.700172] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0769d154-962d-4e2e-bff4-62b8efec0423 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.739395] env[61768]: DEBUG nova.compute.manager [req-0543b15c-0251-4a0c-8f7a-5b1df1bf6178 req-4c2ab7ce-3b36-49a5-bc74-7b724f158bf9 service nova] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Detach interface failed, port_id=c87ca0a3-892f-4fdf-b1d3-a2ff4d3bec10, reason: Instance a36b542c-334f-46b1-9030-ee5c9f8c55c1 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 816.865593] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "d928f14c-b94a-4abf-a053-51015f3bc6c5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.865868] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.866087] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "d928f14c-b94a-4abf-a053-51015f3bc6c5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.866326] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.866657] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.868673] env[61768]: INFO nova.compute.manager [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Terminating instance [ 816.873081] env[61768]: DEBUG nova.compute.manager [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 816.873255] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 816.874138] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-867af13e-c770-4887-a5a4-6ee240275186 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.882854] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 816.883767] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95881819-502e-4769-94ce-97b844c79232 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.892477] env[61768]: DEBUG oslo_vmware.api [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 816.892477] env[61768]: value = "task-1228917" [ 816.892477] env[61768]: _type = "Task" [ 816.892477] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.903353] env[61768]: DEBUG oslo_vmware.api [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228917, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.931142] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Successfully created port: a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.959915] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de5d56d-50d9-45cb-aed1-6fb9429696ee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.970146] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a936e60e-4690-43df-84c5-6ff2d421f0bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.005636] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb90d472-ed1a-4b25-9f71-1bb119eded45 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.015204] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b594ab-36e5-405e-a6d1-ef809881b548 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.031430] env[61768]: DEBUG nova.compute.provider_tree [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.058978] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.082313] env[61768]: INFO nova.compute.manager [-] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Took 1.45 seconds to deallocate network for instance. [ 817.165857] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228916, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074941} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.166340] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.167299] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9976291-6ac4-480b-af30-57acb31daaf6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.195612] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] d4a765bd-a56d-44b7-8db3-c081832d58c9/d4a765bd-a56d-44b7-8db3-c081832d58c9.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.196316] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fef3aa95-bb2f-4638-9a02-1223760c0136 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.223922] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 817.223922] env[61768]: value = "task-1228918" [ 817.223922] env[61768]: _type = "Task" [ 817.223922] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.235271] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228918, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.413074] env[61768]: DEBUG oslo_vmware.api [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228917, 'name': PowerOffVM_Task, 'duration_secs': 0.309158} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.413074] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 817.413175] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 817.413420] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f251d62f-443f-4264-bab6-2273c7843b3e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.494053] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 817.494468] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 817.494788] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleting the datastore file [datastore2] d928f14c-b94a-4abf-a053-51015f3bc6c5 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 817.494788] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-319fa962-251a-4932-9f71-70be8b95688c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.505159] env[61768]: DEBUG oslo_vmware.api [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 817.505159] env[61768]: value = "task-1228920" [ 817.505159] env[61768]: _type = "Task" [ 817.505159] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.515045] env[61768]: DEBUG oslo_vmware.api [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.535094] env[61768]: DEBUG nova.scheduler.client.report [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.587933] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.735708] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228918, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.019713] env[61768]: DEBUG oslo_vmware.api [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.4493} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.020170] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 818.020545] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 818.020761] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 818.021064] env[61768]: INFO nova.compute.manager [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 818.021347] env[61768]: DEBUG oslo.service.loopingcall [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 818.021560] env[61768]: DEBUG nova.compute.manager [-] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.021654] env[61768]: DEBUG nova.network.neutron [-] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 818.041108] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.992s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.044055] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.523s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.046286] env[61768]: INFO nova.compute.claims [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 818.068590] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.081790] env[61768]: INFO nova.scheduler.client.report [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Deleted allocations for instance fbb8c9db-e823-444a-84de-b231cf9b8a8a [ 818.111556] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.112010] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.112326] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.112670] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.112934] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.113215] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.113568] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.113854] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.114174] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.114474] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.114780] env[61768]: DEBUG nova.virt.hardware [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.116952] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26909dec-1903-42b2-b047-d1c6e4dc63c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.128364] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83812550-770a-487e-a3e9-01978a4b444d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.235546] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228918, 'name': ReconfigVM_Task, 'duration_secs': 0.55574} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.235856] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Reconfigured VM instance instance-00000049 to attach disk [datastore1] d4a765bd-a56d-44b7-8db3-c081832d58c9/d4a765bd-a56d-44b7-8db3-c081832d58c9.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.236582] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4b4dc0a3-521b-41f8-b22a-5a8a7cc7401e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.243986] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 818.243986] env[61768]: value = "task-1228921" [ 818.243986] env[61768]: _type = "Task" [ 818.243986] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.252978] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228921, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.350853] env[61768]: DEBUG nova.compute.manager [req-84a8ea49-4b5c-4d84-99e0-b21f5c05a2fe req-5cb1ab82-a8f2-438b-82d2-67b1593c91ff service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Received event network-vif-deleted-2889bdd9-2069-495a-b80a-b20bbec5ef00 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.351083] env[61768]: INFO nova.compute.manager [req-84a8ea49-4b5c-4d84-99e0-b21f5c05a2fe req-5cb1ab82-a8f2-438b-82d2-67b1593c91ff service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Neutron deleted interface 2889bdd9-2069-495a-b80a-b20bbec5ef00; detaching it from the instance and deleting it from the info cache [ 818.351269] env[61768]: DEBUG nova.network.neutron [req-84a8ea49-4b5c-4d84-99e0-b21f5c05a2fe req-5cb1ab82-a8f2-438b-82d2-67b1593c91ff service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.605899] env[61768]: DEBUG oslo_concurrency.lockutils [None req-22758401-7b33-4c06-ad5d-bc954d0e80be tempest-ImagesNegativeTestJSON-1724305828 tempest-ImagesNegativeTestJSON-1724305828-project-member] Lock "fbb8c9db-e823-444a-84de-b231cf9b8a8a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.784s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.754743] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228921, 'name': Rename_Task, 'duration_secs': 0.473885} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.755535] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 818.755922] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c8bdf701-5f83-47d8-bbf8-1abfccbda170 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.766017] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 818.766017] env[61768]: value = "task-1228922" [ 818.766017] env[61768]: _type = "Task" [ 818.766017] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.772694] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228922, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.826214] env[61768]: DEBUG nova.network.neutron [-] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.854271] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1fb80848-bdbd-4524-a5c7-cc3771dde685 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.858859] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Successfully updated port: a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.865304] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c089356-4403-436b-9aff-36870c452460 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.898903] env[61768]: DEBUG nova.compute.manager [req-84a8ea49-4b5c-4d84-99e0-b21f5c05a2fe req-5cb1ab82-a8f2-438b-82d2-67b1593c91ff service nova] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Detach interface failed, port_id=2889bdd9-2069-495a-b80a-b20bbec5ef00, reason: Instance d928f14c-b94a-4abf-a053-51015f3bc6c5 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 819.276637] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228922, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.329557] env[61768]: INFO nova.compute.manager [-] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Took 1.31 seconds to deallocate network for instance. [ 819.366674] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "refresh_cache-3aa722e5-0818-40ae-8220-223d920a7386" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.366674] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "refresh_cache-3aa722e5-0818-40ae-8220-223d920a7386" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.366674] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 819.391730] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-699f9278-4cd8-4ad5-90dd-e83f61e77ec0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.401263] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee595166-0e99-4db4-9238-6d95de2f4335 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.436649] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfc59f93-80e5-499d-9314-79307bda1e43 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.445015] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0a897e-e1cd-47c6-be5f-7c39b7710933 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.460603] env[61768]: DEBUG nova.compute.provider_tree [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.779489] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228922, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.837076] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.917137] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.964172] env[61768]: DEBUG nova.scheduler.client.report [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.147068] env[61768]: DEBUG nova.network.neutron [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Updating instance_info_cache with network_info: [{"id": "a871f4bb-fd0b-4b36-a75e-6ff27d6aa805", "address": "fa:16:3e:07:06:1c", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa871f4bb-fd", "ovs_interfaceid": "a871f4bb-fd0b-4b36-a75e-6ff27d6aa805", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.284255] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228922, 'name': PowerOnVM_Task, 'duration_secs': 1.121225} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.285056] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 820.285478] env[61768]: INFO nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Took 9.61 seconds to spawn the instance on the hypervisor. [ 820.285801] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 820.286719] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29cabd0c-ae6c-4abf-b981-e11a44369e86 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.399360] env[61768]: DEBUG nova.compute.manager [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Received event network-vif-plugged-a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.399607] env[61768]: DEBUG oslo_concurrency.lockutils [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] Acquiring lock "3aa722e5-0818-40ae-8220-223d920a7386-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.399837] env[61768]: DEBUG oslo_concurrency.lockutils [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] Lock "3aa722e5-0818-40ae-8220-223d920a7386-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.400018] env[61768]: DEBUG oslo_concurrency.lockutils [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] Lock "3aa722e5-0818-40ae-8220-223d920a7386-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.400630] env[61768]: DEBUG nova.compute.manager [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] No waiting events found dispatching network-vif-plugged-a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.400842] env[61768]: WARNING nova.compute.manager [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Received unexpected event network-vif-plugged-a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 for instance with vm_state building and task_state spawning. [ 820.401161] env[61768]: DEBUG nova.compute.manager [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Received event network-changed-a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.401259] env[61768]: DEBUG nova.compute.manager [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Refreshing instance network info cache due to event network-changed-a871f4bb-fd0b-4b36-a75e-6ff27d6aa805. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 820.401820] env[61768]: DEBUG oslo_concurrency.lockutils [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] Acquiring lock "refresh_cache-3aa722e5-0818-40ae-8220-223d920a7386" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.469415] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.425s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.469948] env[61768]: DEBUG nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 820.475086] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.784s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.475331] env[61768]: DEBUG nova.objects.instance [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'resources' on Instance uuid 12a543c2-1081-49d7-800b-07f0a2516904 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 820.649064] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "refresh_cache-3aa722e5-0818-40ae-8220-223d920a7386" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.649466] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Instance network_info: |[{"id": "a871f4bb-fd0b-4b36-a75e-6ff27d6aa805", "address": "fa:16:3e:07:06:1c", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa871f4bb-fd", "ovs_interfaceid": "a871f4bb-fd0b-4b36-a75e-6ff27d6aa805", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.649782] env[61768]: DEBUG oslo_concurrency.lockutils [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] Acquired lock "refresh_cache-3aa722e5-0818-40ae-8220-223d920a7386" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.649988] env[61768]: DEBUG nova.network.neutron [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Refreshing network info cache for port a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 820.651285] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:06:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e31a7f15-a808-4199-9071-31fd05e316ea', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a871f4bb-fd0b-4b36-a75e-6ff27d6aa805', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.664948] env[61768]: DEBUG oslo.service.loopingcall [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.668870] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 820.672306] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-410b631d-932e-4b86-9990-9c21e39a4e14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.700325] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 820.700325] env[61768]: value = "task-1228923" [ 820.700325] env[61768]: _type = "Task" [ 820.700325] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.711906] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228923, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.816293] env[61768]: INFO nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Took 33.47 seconds to build instance. [ 820.982250] env[61768]: DEBUG nova.compute.utils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.992611] env[61768]: DEBUG nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.992611] env[61768]: DEBUG nova.network.neutron [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 821.138681] env[61768]: DEBUG nova.policy [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fd5852bc16429aa2904b4351c8249d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7857b5217cc499ca8c9c3b8a5471db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 821.143846] env[61768]: DEBUG nova.network.neutron [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Updated VIF entry in instance network info cache for port a871f4bb-fd0b-4b36-a75e-6ff27d6aa805. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 821.144223] env[61768]: DEBUG nova.network.neutron [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Updating instance_info_cache with network_info: [{"id": "a871f4bb-fd0b-4b36-a75e-6ff27d6aa805", "address": "fa:16:3e:07:06:1c", "network": {"id": "dc8be1fc-cf16-40f1-8fd2-5ef856967a03", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-1408065519-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "54b7f349107b4f56960cae84edbea9e0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e31a7f15-a808-4199-9071-31fd05e316ea", "external-id": "nsx-vlan-transportzone-388", "segmentation_id": 388, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa871f4bb-fd", "ovs_interfaceid": "a871f4bb-fd0b-4b36-a75e-6ff27d6aa805", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.216862] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228923, 'name': CreateVM_Task, 'duration_secs': 0.356292} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.219736] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 821.221445] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.221445] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.221805] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.222181] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-177a2bbd-fda9-4bae-aa2b-0c5249283b99 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.230725] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 821.230725] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b9c47a-4183-de68-2939-7ce6a3f3bc94" [ 821.230725] env[61768]: _type = "Task" [ 821.230725] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.251226] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b9c47a-4183-de68-2939-7ce6a3f3bc94, 'name': SearchDatastore_Task, 'duration_secs': 0.011371} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.252376] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.252686] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 821.252984] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.253231] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.253836] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 821.254197] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d69dc03a-9548-4c8e-a143-3d75c41d0939 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.268734] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 821.268734] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 821.268923] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39adb708-cb2b-44e8-82a3-ce2bcf6483a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.274978] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 821.274978] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e2a59b-1956-2f02-8f5c-e7036bed902c" [ 821.274978] env[61768]: _type = "Task" [ 821.274978] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.290751] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e2a59b-1956-2f02-8f5c-e7036bed902c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.318391] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.803s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.363321] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815dd59c-56ca-4bc5-beea-0b13adf891c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.372143] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5079a71b-a717-4e2a-a764-308e935bb1ba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.418261] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3936a12-b974-41a7-9779-f65bc3ad54ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.427946] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60e875b9-6f93-43ff-b42a-8e286dd76063 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.445219] env[61768]: DEBUG nova.compute.provider_tree [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.488119] env[61768]: DEBUG nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 821.616965] env[61768]: DEBUG nova.network.neutron [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Successfully created port: 43cc41b6-c86c-4fce-a565-3585c3014f7e {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 821.647835] env[61768]: DEBUG oslo_concurrency.lockutils [req-91bd5a27-91b3-4474-921b-0f4b485c2f54 req-adecdbcc-8486-4758-8de3-9e973faef829 service nova] Releasing lock "refresh_cache-3aa722e5-0818-40ae-8220-223d920a7386" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.787853] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e2a59b-1956-2f02-8f5c-e7036bed902c, 'name': SearchDatastore_Task, 'duration_secs': 0.022086} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.788604] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aed7b925-da2b-41b0-8fc6-03392191efa7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.795666] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 821.795666] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52990b83-819c-fc76-0f5b-9e5662f35096" [ 821.795666] env[61768]: _type = "Task" [ 821.795666] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.805497] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52990b83-819c-fc76-0f5b-9e5662f35096, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.948277] env[61768]: DEBUG nova.scheduler.client.report [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 822.308725] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52990b83-819c-fc76-0f5b-9e5662f35096, 'name': SearchDatastore_Task, 'duration_secs': 0.010204} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.309152] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.309568] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 3aa722e5-0818-40ae-8220-223d920a7386/3aa722e5-0818-40ae-8220-223d920a7386.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 822.309865] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8463fd04-6184-482b-ba2e-a4fa87b1165f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.317866] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 822.317866] env[61768]: value = "task-1228924" [ 822.317866] env[61768]: _type = "Task" [ 822.317866] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.327552] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228924, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.453808] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.979s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.457777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.861s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.459765] env[61768]: INFO nova.compute.claims [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.502911] env[61768]: INFO nova.scheduler.client.report [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleted allocations for instance 12a543c2-1081-49d7-800b-07f0a2516904 [ 822.508863] env[61768]: DEBUG nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 822.546787] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 822.546787] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 822.547256] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 822.547540] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 822.547847] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 822.548122] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 822.548599] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 822.548788] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 822.549141] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 822.549445] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 822.549728] env[61768]: DEBUG nova.virt.hardware [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.551740] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f65f9c7-e3fb-4a0d-a208-6a2029576d41 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.567205] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d8f1639-38c9-4860-9d81-41650cb1efb9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.830453] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228924, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49055} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.830772] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 3aa722e5-0818-40ae-8220-223d920a7386/3aa722e5-0818-40ae-8220-223d920a7386.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 822.831024] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 822.831316] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-54e46906-4008-4fb5-a896-4f4707d3cba8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.840120] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 822.840120] env[61768]: value = "task-1228926" [ 822.840120] env[61768]: _type = "Task" [ 822.840120] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.851328] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228926, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.018688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-953c821c-4158-4601-a41d-c299d031cba7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "12a543c2-1081-49d7-800b-07f0a2516904" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.830s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.019696] env[61768]: DEBUG oslo_concurrency.lockutils [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] Acquired lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.020828] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abece3c2-8d1e-4bf5-a3e6-01ddbf1365e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.031499] env[61768]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 823.032252] env[61768]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=61768) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 823.032252] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7df1dbd9-15e0-4f92-8dd4-456621010e23 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.048859] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c11a26-3704-42e9-8228-58f6441e625c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.089185] env[61768]: ERROR root [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-265380' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-265380' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-265380' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-265380'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-265380' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-265380' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-265380'}\n"]: nova.exception.InstanceNotFound: Instance 12a543c2-1081-49d7-800b-07f0a2516904 could not be found. [ 823.089185] env[61768]: DEBUG oslo_concurrency.lockutils [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] Releasing lock "12a543c2-1081-49d7-800b-07f0a2516904" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.089185] env[61768]: DEBUG nova.compute.manager [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Detach interface failed, port_id=a0589e32-427b-4fea-a7f6-6de811fcebdb, reason: Instance 12a543c2-1081-49d7-800b-07f0a2516904 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 823.089185] env[61768]: DEBUG nova.compute.manager [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Received event network-vif-deleted-cb8850cf-e632-419a-86f0-4c532e443213 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.089185] env[61768]: INFO nova.compute.manager [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Neutron deleted interface cb8850cf-e632-419a-86f0-4c532e443213; detaching it from the instance and deleting it from the info cache [ 823.089470] env[61768]: DEBUG nova.network.neutron [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Updating instance_info_cache with network_info: [{"id": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "address": "fa:16:3e:23:2f:fd", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.182", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3a76322f-8d", "ovs_interfaceid": "3a76322f-8de0-44f9-9b42-ae3f6c592d3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.192962] env[61768]: DEBUG nova.compute.manager [req-05b448cb-feb9-4213-8da5-f0dc00196656 req-8bde11fa-ada1-42cd-b61f-91a2b6ac22fb service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Received event network-vif-plugged-43cc41b6-c86c-4fce-a565-3585c3014f7e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 823.193211] env[61768]: DEBUG oslo_concurrency.lockutils [req-05b448cb-feb9-4213-8da5-f0dc00196656 req-8bde11fa-ada1-42cd-b61f-91a2b6ac22fb service nova] Acquiring lock "37b729bf-ad5e-4e17-b11b-77b504c049cd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.193933] env[61768]: DEBUG oslo_concurrency.lockutils [req-05b448cb-feb9-4213-8da5-f0dc00196656 req-8bde11fa-ada1-42cd-b61f-91a2b6ac22fb service nova] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.194146] env[61768]: DEBUG oslo_concurrency.lockutils [req-05b448cb-feb9-4213-8da5-f0dc00196656 req-8bde11fa-ada1-42cd-b61f-91a2b6ac22fb service nova] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.194380] env[61768]: DEBUG nova.compute.manager [req-05b448cb-feb9-4213-8da5-f0dc00196656 req-8bde11fa-ada1-42cd-b61f-91a2b6ac22fb service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] No waiting events found dispatching network-vif-plugged-43cc41b6-c86c-4fce-a565-3585c3014f7e {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 823.194505] env[61768]: WARNING nova.compute.manager [req-05b448cb-feb9-4213-8da5-f0dc00196656 req-8bde11fa-ada1-42cd-b61f-91a2b6ac22fb service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Received unexpected event network-vif-plugged-43cc41b6-c86c-4fce-a565-3585c3014f7e for instance with vm_state building and task_state spawning. [ 823.243261] env[61768]: DEBUG nova.network.neutron [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Successfully updated port: 43cc41b6-c86c-4fce-a565-3585c3014f7e {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 823.350382] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228926, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068477} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.350667] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.351448] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049371bf-5f0d-43d3-b9f9-f4e73a61c03f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.373825] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] 3aa722e5-0818-40ae-8220-223d920a7386/3aa722e5-0818-40ae-8220-223d920a7386.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.374054] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0591416c-863c-4b9d-917c-3f69ba9cc393 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.394499] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 823.394499] env[61768]: value = "task-1228930" [ 823.394499] env[61768]: _type = "Task" [ 823.394499] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.402630] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228930, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.593547] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03c86a9f-064e-4d5e-b9d9-34a6b55951f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.604912] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb5a7d2-6bb5-407b-a8a1-a8410ed0a8e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.641869] env[61768]: DEBUG nova.compute.manager [req-3ff16b94-3eee-4f5d-a805-da1e17241ec1 req-46dad1f2-1b35-4bb9-b035-12978e2794f6 service nova] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Detach interface failed, port_id=cb8850cf-e632-419a-86f0-4c532e443213, reason: Instance 12a543c2-1081-49d7-800b-07f0a2516904 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 823.747191] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96980b30-d57f-408d-8b2f-bc90d1d7b40a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.750828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "refresh_cache-37b729bf-ad5e-4e17-b11b-77b504c049cd" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.750828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "refresh_cache-37b729bf-ad5e-4e17-b11b-77b504c049cd" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.750828] env[61768]: DEBUG nova.network.neutron [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.758078] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c5b014-1706-4366-8f6f-76d8a6087d84 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.792398] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2a9005-b459-498f-8b66-9fc642e02d73 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.801066] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2895339a-7013-42c1-98cc-d0cee0b7e4ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.817066] env[61768]: DEBUG nova.compute.provider_tree [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.906698] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228930, 'name': ReconfigVM_Task, 'duration_secs': 0.31001} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.906995] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Reconfigured VM instance instance-0000004a to attach disk [datastore2] 3aa722e5-0818-40ae-8220-223d920a7386/3aa722e5-0818-40ae-8220-223d920a7386.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 823.907698] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b1dc5e50-fd12-4568-ab2c-25c5287ea7bf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.916238] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 823.916238] env[61768]: value = "task-1228931" [ 823.916238] env[61768]: _type = "Task" [ 823.916238] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.925603] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228931, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.283075] env[61768]: DEBUG nova.network.neutron [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 824.319551] env[61768]: DEBUG nova.scheduler.client.report [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 824.430453] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228931, 'name': Rename_Task, 'duration_secs': 0.182266} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.433320] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 824.433611] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09a7e30c-d63d-49f4-843c-af448f39c106 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.442411] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 824.442411] env[61768]: value = "task-1228932" [ 824.442411] env[61768]: _type = "Task" [ 824.442411] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.450989] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228932, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.476750] env[61768]: DEBUG nova.network.neutron [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Updating instance_info_cache with network_info: [{"id": "43cc41b6-c86c-4fce-a565-3585c3014f7e", "address": "fa:16:3e:4a:f4:60", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43cc41b6-c8", "ovs_interfaceid": "43cc41b6-c86c-4fce-a565-3585c3014f7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.738503] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.738820] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.739018] env[61768]: INFO nova.compute.manager [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Shelving [ 824.825824] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.368s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.826576] env[61768]: DEBUG nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 824.829551] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 25.858s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.954209] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228932, 'name': PowerOnVM_Task} progress is 92%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.981245] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "refresh_cache-37b729bf-ad5e-4e17-b11b-77b504c049cd" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.981717] env[61768]: DEBUG nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Instance network_info: |[{"id": "43cc41b6-c86c-4fce-a565-3585c3014f7e", "address": "fa:16:3e:4a:f4:60", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43cc41b6-c8", "ovs_interfaceid": "43cc41b6-c86c-4fce-a565-3585c3014f7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 824.982451] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4a:f4:60', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43cc41b6-c86c-4fce-a565-3585c3014f7e', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 824.993397] env[61768]: DEBUG oslo.service.loopingcall [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.993696] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 824.994199] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3e7aac4-2840-44cc-b36d-a5978a11b508 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.017394] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 825.017394] env[61768]: value = "task-1228933" [ 825.017394] env[61768]: _type = "Task" [ 825.017394] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.027085] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228933, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.249820] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.250184] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a29fea55-f6df-411a-81c6-3d82d81a11ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.259418] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 825.259418] env[61768]: value = "task-1228934" [ 825.259418] env[61768]: _type = "Task" [ 825.259418] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.271069] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228934, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.277839] env[61768]: DEBUG nova.compute.manager [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Received event network-changed-43cc41b6-c86c-4fce-a565-3585c3014f7e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.278459] env[61768]: DEBUG nova.compute.manager [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Refreshing instance network info cache due to event network-changed-43cc41b6-c86c-4fce-a565-3585c3014f7e. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.278459] env[61768]: DEBUG oslo_concurrency.lockutils [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] Acquiring lock "refresh_cache-37b729bf-ad5e-4e17-b11b-77b504c049cd" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.278687] env[61768]: DEBUG oslo_concurrency.lockutils [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] Acquired lock "refresh_cache-37b729bf-ad5e-4e17-b11b-77b504c049cd" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.278735] env[61768]: DEBUG nova.network.neutron [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Refreshing network info cache for port 43cc41b6-c86c-4fce-a565-3585c3014f7e {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 825.335086] env[61768]: INFO nova.compute.claims [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.340022] env[61768]: DEBUG nova.compute.utils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 825.344015] env[61768]: DEBUG nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 825.344015] env[61768]: DEBUG nova.network.neutron [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 825.399233] env[61768]: DEBUG nova.policy [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2280e9e6bde740c3aca30ebf70c7f73d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc444d04ff864300a45a55a3c67507e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 825.455638] env[61768]: DEBUG oslo_vmware.api [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228932, 'name': PowerOnVM_Task, 'duration_secs': 0.568205} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.455918] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 825.456158] env[61768]: INFO nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Took 7.39 seconds to spawn the instance on the hypervisor. [ 825.456345] env[61768]: DEBUG nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.457184] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bf79ff-df56-4582-8997-c1848eaa5412 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.533017] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228933, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.771188] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228934, 'name': PowerOffVM_Task, 'duration_secs': 0.327478} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.771484] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.772461] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9c4a32-7cda-44c2-bd4e-d1358ffa3f58 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.797645] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c8cf61-1b0b-470e-acd9-b801789b24cd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.844851] env[61768]: INFO nova.compute.resource_tracker [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating resource usage from migration 05095ce1-1d9e-4bdc-8f8d-2b88b84020e3 [ 825.849052] env[61768]: DEBUG nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.854608] env[61768]: DEBUG nova.network.neutron [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Successfully created port: e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.977043] env[61768]: INFO nova.compute.manager [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Took 35.95 seconds to build instance. [ 825.989218] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "ecd08348-b22f-49c5-b13f-b187506e38d6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.989502] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.033968] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228933, 'name': CreateVM_Task, 'duration_secs': 0.739481} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.035290] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 826.036752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.036752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.036752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 826.036953] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68f04f23-4a90-4271-9afb-4898113c4890 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.043959] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 826.043959] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52331f37-686d-c297-32d8-eaf0a2d84834" [ 826.043959] env[61768]: _type = "Task" [ 826.043959] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.058485] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52331f37-686d-c297-32d8-eaf0a2d84834, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.202752] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac8b1ca3-a8c7-4ebe-a119-d12481a680c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.216045] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf27e8c-1aec-48f0-87a0-e0a46785feae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.219380] env[61768]: DEBUG nova.network.neutron [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Updated VIF entry in instance network info cache for port 43cc41b6-c86c-4fce-a565-3585c3014f7e. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 826.219778] env[61768]: DEBUG nova.network.neutron [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Updating instance_info_cache with network_info: [{"id": "43cc41b6-c86c-4fce-a565-3585c3014f7e", "address": "fa:16:3e:4a:f4:60", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43cc41b6-c8", "ovs_interfaceid": "43cc41b6-c86c-4fce-a565-3585c3014f7e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.257094] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-843797a5-160b-4f49-90e2-f314cff48c72 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.267827] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70c350b-1622-45dd-b7e7-f04c8cb91930 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.283770] env[61768]: DEBUG nova.compute.provider_tree [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 826.314323] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 826.314648] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-0066888e-d9a0-4e13-9aa5-6acf87625c55 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.317103] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.317315] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.326972] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 826.326972] env[61768]: value = "task-1228936" [ 826.326972] env[61768]: _type = "Task" [ 826.326972] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.340130] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228936, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.479969] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eb961a5b-af7d-460a-91e9-505e6811a9cc tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "3aa722e5-0818-40ae-8220-223d920a7386" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.935s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.493707] env[61768]: DEBUG nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.554538] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52331f37-686d-c297-32d8-eaf0a2d84834, 'name': SearchDatastore_Task, 'duration_secs': 0.022458} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.554847] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.555101] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 826.555357] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.555514] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.556121] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 826.556121] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f05f8b3d-1784-4097-94f9-700c78def144 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.572011] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 826.572212] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 826.572954] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab4bfa60-dbda-40b4-9ec5-7deeeb174561 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.578408] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 826.578408] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52053d47-0a3f-d4df-1d4a-33faadd562c6" [ 826.578408] env[61768]: _type = "Task" [ 826.578408] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.586224] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52053d47-0a3f-d4df-1d4a-33faadd562c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.607087] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "d4a765bd-a56d-44b7-8db3-c081832d58c9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.607315] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.607506] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "d4a765bd-a56d-44b7-8db3-c081832d58c9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.607691] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.607863] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.610037] env[61768]: INFO nova.compute.manager [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Terminating instance [ 826.611793] env[61768]: DEBUG nova.compute.manager [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 826.611992] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 826.612762] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb669e10-f0d8-43f6-808c-58dbdbfe802f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.622037] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 826.622271] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6939a0c6-374a-4890-ab14-b1611b14e03d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.629067] env[61768]: DEBUG oslo_vmware.api [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 826.629067] env[61768]: value = "task-1228937" [ 826.629067] env[61768]: _type = "Task" [ 826.629067] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.636742] env[61768]: DEBUG oslo_vmware.api [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228937, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.679267] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "3aa722e5-0818-40ae-8220-223d920a7386" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.680873] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "3aa722e5-0818-40ae-8220-223d920a7386" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.680873] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "3aa722e5-0818-40ae-8220-223d920a7386-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.680873] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "3aa722e5-0818-40ae-8220-223d920a7386-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.680873] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "3aa722e5-0818-40ae-8220-223d920a7386-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.682544] env[61768]: INFO nova.compute.manager [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Terminating instance [ 826.684430] env[61768]: DEBUG nova.compute.manager [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 826.684630] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 826.685498] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976798eb-c127-410b-80f1-a7ee9ca905e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.694593] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 826.694978] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1a25ed30-317d-44b4-a7ca-910b5049b970 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.702535] env[61768]: DEBUG oslo_vmware.api [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 826.702535] env[61768]: value = "task-1228938" [ 826.702535] env[61768]: _type = "Task" [ 826.702535] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.711881] env[61768]: DEBUG oslo_vmware.api [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228938, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.723252] env[61768]: DEBUG oslo_concurrency.lockutils [req-4ed0d673-6f9a-48bd-8995-7eb26e0fd85c req-33975400-e08d-434b-87e7-67f092761768 service nova] Releasing lock "refresh_cache-37b729bf-ad5e-4e17-b11b-77b504c049cd" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.787057] env[61768]: DEBUG nova.scheduler.client.report [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.825053] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 826.825229] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 826.839510] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228936, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.865148] env[61768]: DEBUG nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.896535] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.896804] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.896971] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.897172] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.897329] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.897485] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.897702] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.897870] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.898174] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.898459] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.898704] env[61768]: DEBUG nova.virt.hardware [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.899682] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b530865-4237-42df-9074-fcaf0daa566c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.908916] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b19b69-680a-4a24-872d-4382f30c0970 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.015508] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.089276] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52053d47-0a3f-d4df-1d4a-33faadd562c6, 'name': SearchDatastore_Task, 'duration_secs': 0.062268} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.090173] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b356f244-2c4a-4d49-9c76-ffb9e0c8d423 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.095616] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 827.095616] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5257574d-72d5-f496-8637-ff3966a99be6" [ 827.095616] env[61768]: _type = "Task" [ 827.095616] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.103275] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5257574d-72d5-f496-8637-ff3966a99be6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.137854] env[61768]: DEBUG oslo_vmware.api [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228937, 'name': PowerOffVM_Task, 'duration_secs': 0.233818} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.138151] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 827.138493] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 827.138804] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-090b9300-5d5f-4e38-8a17-56e1dad08150 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.213656] env[61768]: DEBUG oslo_vmware.api [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228938, 'name': PowerOffVM_Task, 'duration_secs': 0.206317} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.214988] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 827.215217] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 827.215513] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 827.215719] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 827.215901] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleting the datastore file [datastore1] d4a765bd-a56d-44b7-8db3-c081832d58c9 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.216238] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e176ab82-544c-4b76-bfab-9f6b9156a30c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.217839] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a8de959d-437b-4257-8d39-8456a99ddca1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.225835] env[61768]: DEBUG oslo_vmware.api [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 827.225835] env[61768]: value = "task-1228941" [ 827.225835] env[61768]: _type = "Task" [ 827.225835] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.234371] env[61768]: DEBUG oslo_vmware.api [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228941, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.292719] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.463s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.292931] env[61768]: INFO nova.compute.manager [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Migrating [ 827.293211] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.293416] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.296798] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.850s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.298688] env[61768]: INFO nova.compute.claims [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 827.301872] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 827.302473] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 827.302723] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleting the datastore file [datastore2] 3aa722e5-0818-40ae-8220-223d920a7386 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 827.304128] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-334fe64c-5fc3-4e5f-b53b-f14f0bb90030 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.311359] env[61768]: DEBUG oslo_vmware.api [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for the task: (returnval){ [ 827.311359] env[61768]: value = "task-1228942" [ 827.311359] env[61768]: _type = "Task" [ 827.311359] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.320153] env[61768]: DEBUG oslo_vmware.api [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228942, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.337804] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Didn't find any instances for network info cache update. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 827.338091] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228936, 'name': CreateSnapshot_Task, 'duration_secs': 0.846513} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.338311] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.338568] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 827.338815] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.339560] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a343cdc-84e7-4cef-87ee-ca4cab08d48f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.341954] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.342442] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.343009] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.343223] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.343379] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 827.343543] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 827.606975] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5257574d-72d5-f496-8637-ff3966a99be6, 'name': SearchDatastore_Task, 'duration_secs': 0.011426} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.607784] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.608201] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 37b729bf-ad5e-4e17-b11b-77b504c049cd/37b729bf-ad5e-4e17-b11b-77b504c049cd.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 827.608444] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d5921cb1-c9f8-4c5f-94e7-30d9142e9205 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.616582] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 827.616582] env[61768]: value = "task-1228943" [ 827.616582] env[61768]: _type = "Task" [ 827.616582] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.627782] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.646941] env[61768]: DEBUG nova.compute.manager [req-79a1753a-dc35-4d4a-af95-0ac8b2019a79 req-05a460e4-7b66-4563-b72a-7d73bc32350a service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Received event network-vif-plugged-e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.647319] env[61768]: DEBUG oslo_concurrency.lockutils [req-79a1753a-dc35-4d4a-af95-0ac8b2019a79 req-05a460e4-7b66-4563-b72a-7d73bc32350a service nova] Acquiring lock "8e0e13a7-eac5-4176-8536-2906d13c390e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.647480] env[61768]: DEBUG oslo_concurrency.lockutils [req-79a1753a-dc35-4d4a-af95-0ac8b2019a79 req-05a460e4-7b66-4563-b72a-7d73bc32350a service nova] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.647718] env[61768]: DEBUG oslo_concurrency.lockutils [req-79a1753a-dc35-4d4a-af95-0ac8b2019a79 req-05a460e4-7b66-4563-b72a-7d73bc32350a service nova] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.647961] env[61768]: DEBUG nova.compute.manager [req-79a1753a-dc35-4d4a-af95-0ac8b2019a79 req-05a460e4-7b66-4563-b72a-7d73bc32350a service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] No waiting events found dispatching network-vif-plugged-e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 827.648216] env[61768]: WARNING nova.compute.manager [req-79a1753a-dc35-4d4a-af95-0ac8b2019a79 req-05a460e4-7b66-4563-b72a-7d73bc32350a service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Received unexpected event network-vif-plugged-e02bb27b-f94f-4f93-859a-5af2646daaaa for instance with vm_state building and task_state spawning. [ 827.736446] env[61768]: DEBUG oslo_vmware.api [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228941, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149082} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.736782] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.736983] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 827.737196] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 827.737384] env[61768]: INFO nova.compute.manager [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 827.737646] env[61768]: DEBUG oslo.service.loopingcall [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.737846] env[61768]: DEBUG nova.compute.manager [-] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.737938] env[61768]: DEBUG nova.network.neutron [-] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.809186] env[61768]: INFO nova.compute.rpcapi [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 827.811455] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.844545] env[61768]: DEBUG oslo_vmware.api [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Task: {'id': task-1228942, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151021} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.844712] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 827.845149] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 827.845462] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 827.845736] env[61768]: INFO nova.compute.manager [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Took 1.16 seconds to destroy the instance on the hypervisor. [ 827.846097] env[61768]: DEBUG oslo.service.loopingcall [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.847073] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.847423] env[61768]: DEBUG nova.compute.manager [-] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.847528] env[61768]: DEBUG nova.network.neutron [-] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.858915] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 827.860534] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-87dd312e-d747-4fe4-9d6d-900a64a18d22 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.871881] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 827.871881] env[61768]: value = "task-1228945" [ 827.871881] env[61768]: _type = "Task" [ 827.871881] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.883136] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228945, 'name': CloneVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.947935] env[61768]: DEBUG nova.network.neutron [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Successfully updated port: e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 828.130046] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228943, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.337147] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.337369] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.337554] env[61768]: DEBUG nova.network.neutron [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.347679] env[61768]: DEBUG nova.compute.manager [req-5f856399-f33a-4d90-83bb-7f24378b2a1e req-689edfbf-7b09-4f32-bccd-d67b3dcfc8f7 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Received event network-vif-deleted-a871f4bb-fd0b-4b36-a75e-6ff27d6aa805 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 828.347679] env[61768]: INFO nova.compute.manager [req-5f856399-f33a-4d90-83bb-7f24378b2a1e req-689edfbf-7b09-4f32-bccd-d67b3dcfc8f7 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Neutron deleted interface a871f4bb-fd0b-4b36-a75e-6ff27d6aa805; detaching it from the instance and deleting it from the info cache [ 828.348026] env[61768]: DEBUG nova.network.neutron [req-5f856399-f33a-4d90-83bb-7f24378b2a1e req-689edfbf-7b09-4f32-bccd-d67b3dcfc8f7 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.382544] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228945, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.452564] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.452735] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.452890] env[61768]: DEBUG nova.network.neutron [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.591157] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efb06858-497f-4b11-86a0-6084e225bb91 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.600035] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a23099f9-226d-4d53-95d4-765b5471af4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.638814] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2ca7612-6027-4a78-81e6-50204ff12cb3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.650276] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ab1bf5-2dd6-4e07-b218-7052546c5639 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.654120] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.561594} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.654391] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 37b729bf-ad5e-4e17-b11b-77b504c049cd/37b729bf-ad5e-4e17-b11b-77b504c049cd.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 828.654619] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 828.655248] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0557c6de-b9a7-411b-b2e2-91a3e4212138 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.667906] env[61768]: DEBUG nova.compute.provider_tree [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.670673] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 828.670673] env[61768]: value = "task-1228946" [ 828.670673] env[61768]: _type = "Task" [ 828.670673] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.680826] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228946, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.739623] env[61768]: DEBUG nova.network.neutron [-] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.784023] env[61768]: DEBUG nova.network.neutron [-] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.850372] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-de1a4b07-629b-4ee9-9fb8-1e2c4a893a71 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.861148] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8be21122-f4be-4a2d-b41a-db186aae646f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.881089] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228945, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.893597] env[61768]: DEBUG nova.compute.manager [req-5f856399-f33a-4d90-83bb-7f24378b2a1e req-689edfbf-7b09-4f32-bccd-d67b3dcfc8f7 service nova] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Detach interface failed, port_id=a871f4bb-fd0b-4b36-a75e-6ff27d6aa805, reason: Instance 3aa722e5-0818-40ae-8220-223d920a7386 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 828.984045] env[61768]: DEBUG nova.network.neutron [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.051120] env[61768]: DEBUG nova.network.neutron [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance_info_cache with network_info: [{"id": "0e23d603-ab56-4c07-b562-335a39497bac", "address": "fa:16:3e:3d:a9:43", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e23d603-ab", "ovs_interfaceid": "0e23d603-ab56-4c07-b562-335a39497bac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.173300] env[61768]: DEBUG nova.scheduler.client.report [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 829.186453] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228946, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076785} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.186827] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 829.187836] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ca4e5f-3bcb-4a83-9483-8583834176e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.193033] env[61768]: DEBUG nova.network.neutron [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Updating instance_info_cache with network_info: [{"id": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "address": "fa:16:3e:39:0c:b3", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02bb27b-f9", "ovs_interfaceid": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.212732] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 37b729bf-ad5e-4e17-b11b-77b504c049cd/37b729bf-ad5e-4e17-b11b-77b504c049cd.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 829.213788] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f770bf35-e3ed-4534-85bf-e36e1c605026 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.235574] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 829.235574] env[61768]: value = "task-1228947" [ 829.235574] env[61768]: _type = "Task" [ 829.235574] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.242090] env[61768]: INFO nova.compute.manager [-] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Took 1.50 seconds to deallocate network for instance. [ 829.251229] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228947, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.286707] env[61768]: INFO nova.compute.manager [-] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Took 1.44 seconds to deallocate network for instance. [ 829.381841] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228945, 'name': CloneVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.554188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.681241] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.681888] env[61768]: DEBUG nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.684479] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.835s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.685896] env[61768]: INFO nova.compute.claims [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.694912] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.695195] env[61768]: DEBUG nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Instance network_info: |[{"id": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "address": "fa:16:3e:39:0c:b3", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02bb27b-f9", "ovs_interfaceid": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 829.695769] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:39:0c:b3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13e71dbb-4279-427c-b39d-ba5df9895e58', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e02bb27b-f94f-4f93-859a-5af2646daaaa', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 829.703758] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Creating folder: Project (fc444d04ff864300a45a55a3c67507e5). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.704153] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f637e254-c6e2-4ed2-b5bc-b216f9c57057 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.718106] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Created folder: Project (fc444d04ff864300a45a55a3c67507e5) in parent group-v265360. [ 829.718417] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Creating folder: Instances. Parent ref: group-v265466. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 829.718964] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-55489d61-dd95-4e95-85de-0d5ce9c4ed54 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.729562] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Created folder: Instances in parent group-v265466. [ 829.729830] env[61768]: DEBUG oslo.service.loopingcall [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.730097] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 829.730248] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e213f2ee-27bc-4d66-b2b5-bc558df73a50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.755858] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.756178] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228947, 'name': ReconfigVM_Task, 'duration_secs': 0.397068} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.757471] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 37b729bf-ad5e-4e17-b11b-77b504c049cd/37b729bf-ad5e-4e17-b11b-77b504c049cd.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 829.758137] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 829.758137] env[61768]: value = "task-1228950" [ 829.758137] env[61768]: _type = "Task" [ 829.758137] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.758375] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2bd6abf5-9e12-49a1-8416-3348cf41541e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.768616] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228950, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.769911] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 829.769911] env[61768]: value = "task-1228951" [ 829.769911] env[61768]: _type = "Task" [ 829.769911] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.778091] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228951, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.793514] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.800171] env[61768]: DEBUG nova.compute.manager [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Received event network-changed-e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.800401] env[61768]: DEBUG nova.compute.manager [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Refreshing instance network info cache due to event network-changed-e02bb27b-f94f-4f93-859a-5af2646daaaa. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.800672] env[61768]: DEBUG oslo_concurrency.lockutils [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] Acquiring lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.800871] env[61768]: DEBUG oslo_concurrency.lockutils [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] Acquired lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.801080] env[61768]: DEBUG nova.network.neutron [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Refreshing network info cache for port e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 829.883188] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228945, 'name': CloneVM_Task, 'duration_secs': 1.561864} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.883488] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Created linked-clone VM from snapshot [ 829.884270] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef772927-12ad-4064-aa70-5927a62fdfc5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.893197] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Uploading image 1e83b86d-089f-4b79-940c-ef76d051d4f3 {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 829.923760] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 829.923760] env[61768]: value = "vm-265465" [ 829.923760] env[61768]: _type = "VirtualMachine" [ 829.923760] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 829.924105] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-056cf172-3455-40e9-9803-37023573bd8f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.932226] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease: (returnval){ [ 829.932226] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e9050e-548f-483d-f9b4-59e0456dba74" [ 829.932226] env[61768]: _type = "HttpNfcLease" [ 829.932226] env[61768]: } obtained for exporting VM: (result){ [ 829.932226] env[61768]: value = "vm-265465" [ 829.932226] env[61768]: _type = "VirtualMachine" [ 829.932226] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 829.932571] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the lease: (returnval){ [ 829.932571] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e9050e-548f-483d-f9b4-59e0456dba74" [ 829.932571] env[61768]: _type = "HttpNfcLease" [ 829.932571] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 829.939956] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 829.939956] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e9050e-548f-483d-f9b4-59e0456dba74" [ 829.939956] env[61768]: _type = "HttpNfcLease" [ 829.939956] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 830.190205] env[61768]: DEBUG nova.compute.utils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.193836] env[61768]: DEBUG nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.194123] env[61768]: DEBUG nova.network.neutron [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 830.267385] env[61768]: DEBUG nova.policy [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd1aa7c174648888b58f6a6df6f82a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfce0cf4a5d14042a929dac65aeb2ae3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 830.275987] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228950, 'name': CreateVM_Task, 'duration_secs': 0.378665} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.281311] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 830.282483] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.282834] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.283673] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 830.284552] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0f899f3f-8208-4811-8351-a21c1fb7ddd0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.290493] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228951, 'name': Rename_Task, 'duration_secs': 0.156002} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.291104] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 830.291380] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be3ca19e-3ca2-4b79-afb3-238d17b94648 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.293947] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 830.293947] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525a2c35-dcb8-0cf3-033d-03053cd61f1a" [ 830.293947] env[61768]: _type = "Task" [ 830.293947] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.299354] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 830.299354] env[61768]: value = "task-1228954" [ 830.299354] env[61768]: _type = "Task" [ 830.299354] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.302864] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525a2c35-dcb8-0cf3-033d-03053cd61f1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.313558] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228954, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.440873] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 830.440873] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e9050e-548f-483d-f9b4-59e0456dba74" [ 830.440873] env[61768]: _type = "HttpNfcLease" [ 830.440873] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 830.441263] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 830.441263] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e9050e-548f-483d-f9b4-59e0456dba74" [ 830.441263] env[61768]: _type = "HttpNfcLease" [ 830.441263] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 830.441974] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ef5bde9-dcb6-4415-b21b-9d51b3f73cce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.450155] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52510471-ea17-2ca9-5eab-bf09932be0a9/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 830.450358] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52510471-ea17-2ca9-5eab-bf09932be0a9/disk-0.vmdk for reading. {{(pid=61768) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 830.548084] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-03d754fb-543c-4aea-b17c-e01f30283f0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.615198] env[61768]: DEBUG nova.network.neutron [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Successfully created port: bc255400-bfd6-496f-a447-290139231203 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.617855] env[61768]: DEBUG nova.network.neutron [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Updated VIF entry in instance network info cache for port e02bb27b-f94f-4f93-859a-5af2646daaaa. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 830.618268] env[61768]: DEBUG nova.network.neutron [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Updating instance_info_cache with network_info: [{"id": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "address": "fa:16:3e:39:0c:b3", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02bb27b-f9", "ovs_interfaceid": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.700971] env[61768]: DEBUG nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.811751] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525a2c35-dcb8-0cf3-033d-03053cd61f1a, 'name': SearchDatastore_Task, 'duration_secs': 0.037178} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.813474] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.813474] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 830.813474] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.813474] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.813474] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 830.813782] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e319dd57-7a95-48d3-b9e3-a737a342a31d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.819385] env[61768]: DEBUG oslo_vmware.api [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228954, 'name': PowerOnVM_Task, 'duration_secs': 0.515219} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.822363] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 830.822618] env[61768]: INFO nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Took 8.31 seconds to spawn the instance on the hypervisor. [ 830.822899] env[61768]: DEBUG nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.824261] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f144826-6ce5-4f15-8c73-4e29367e6f60 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.830252] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 830.830460] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 830.832795] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2536f11a-3051-419f-bd13-b206dd5f812c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.849030] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 830.849030] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5263a2cc-1656-6341-f309-a7fc0462ce28" [ 830.849030] env[61768]: _type = "Task" [ 830.849030] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.855381] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5263a2cc-1656-6341-f309-a7fc0462ce28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.057048] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90676bea-5265-4024-b300-d7cdbbf4ec7e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.069828] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a17b32-b18a-40a5-bb27-f9cf5f078243 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.074734] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-962cb784-b899-40bf-9dff-eb1193c021d7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.096033] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance '4f429be7-c594-4a72-aebd-0b746219231c' progress to 0 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 831.127777] env[61768]: DEBUG oslo_concurrency.lockutils [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] Releasing lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.128222] env[61768]: DEBUG nova.compute.manager [req-606664e1-4f2e-497d-9a79-53e9d047c18e req-83acc334-94bb-49eb-be7f-7ff2acbdb680 service nova] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Received event network-vif-deleted-667907df-2454-4a31-9b6f-4e202b308713 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.129856] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55589e6b-7457-4104-a98b-2177c745456f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.143787] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8018cb4-ae23-4e38-a3d5-f5da4ccd67de {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.160565] env[61768]: DEBUG nova.compute.provider_tree [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 831.353231] env[61768]: INFO nova.compute.manager [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Took 37.85 seconds to build instance. [ 831.359107] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5263a2cc-1656-6341-f309-a7fc0462ce28, 'name': SearchDatastore_Task, 'duration_secs': 0.012701} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.360019] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9af30602-0a83-47ab-884a-baf0bc3ecd1a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.367106] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 831.367106] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522b6b7e-f239-9451-b007-be84333827c1" [ 831.367106] env[61768]: _type = "Task" [ 831.367106] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.377534] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522b6b7e-f239-9451-b007-be84333827c1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.632710] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.632710] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-415f4f9d-6a84-4712-afc8-2834419191ba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.644159] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 831.644159] env[61768]: value = "task-1228955" [ 831.644159] env[61768]: _type = "Task" [ 831.644159] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.657638] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.682259] env[61768]: ERROR nova.scheduler.client.report [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [req-e57828fa-ef38-4a55-97c1-314df24ece31] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e57828fa-ef38-4a55-97c1-314df24ece31"}]} [ 831.701053] env[61768]: DEBUG nova.scheduler.client.report [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 831.712046] env[61768]: DEBUG nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.717769] env[61768]: DEBUG nova.scheduler.client.report [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 831.718147] env[61768]: DEBUG nova.compute.provider_tree [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 831.730884] env[61768]: DEBUG nova.scheduler.client.report [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 831.736523] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.737100] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.737468] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.738439] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.738439] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.738439] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.738439] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.738969] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.738969] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.739103] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.739447] env[61768]: DEBUG nova.virt.hardware [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.740394] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9155b1fa-4ea9-450f-b790-fb18c4c4b8c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.750159] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b0c282c-56b0-4ed6-ae91-01350b29d99c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.756665] env[61768]: DEBUG nova.scheduler.client.report [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 831.773688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "37b729bf-ad5e-4e17-b11b-77b504c049cd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.855195] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bf582b2d-d796-4fe7-ba3b-b6b0f81bbec4 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.125s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.855703] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.082s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.855946] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "37b729bf-ad5e-4e17-b11b-77b504c049cd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.856176] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.856352] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.858817] env[61768]: INFO nova.compute.manager [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Terminating instance [ 831.861342] env[61768]: DEBUG nova.compute.manager [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 831.861623] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 831.862711] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fb4473e-1245-447b-8bf2-603d678bb517 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.876610] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 831.877486] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7fddb934-3b84-4601-84f2-7dcc259dd846 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.883466] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522b6b7e-f239-9451-b007-be84333827c1, 'name': SearchDatastore_Task, 'duration_secs': 0.019857} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.887827] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.888284] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 8e0e13a7-eac5-4176-8536-2906d13c390e/8e0e13a7-eac5-4176-8536-2906d13c390e.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 831.889621] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4d91106a-f552-4dd6-a80c-2c4c81a059e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.894244] env[61768]: DEBUG oslo_vmware.api [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 831.894244] env[61768]: value = "task-1228956" [ 831.894244] env[61768]: _type = "Task" [ 831.894244] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.901385] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 831.901385] env[61768]: value = "task-1228957" [ 831.901385] env[61768]: _type = "Task" [ 831.901385] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.916544] env[61768]: DEBUG oslo_vmware.api [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.922537] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228957, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.107434] env[61768]: DEBUG nova.compute.manager [req-65fd4031-dcd3-44b4-a70e-6805317a2dda req-bc20992d-9c12-4c7f-8a47-cdd9b33a9dd9 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Received event network-vif-plugged-bc255400-bfd6-496f-a447-290139231203 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.107751] env[61768]: DEBUG oslo_concurrency.lockutils [req-65fd4031-dcd3-44b4-a70e-6805317a2dda req-bc20992d-9c12-4c7f-8a47-cdd9b33a9dd9 service nova] Acquiring lock "41f6300d-7462-4e11-b32b-8c892e87bafc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.108109] env[61768]: DEBUG oslo_concurrency.lockutils [req-65fd4031-dcd3-44b4-a70e-6805317a2dda req-bc20992d-9c12-4c7f-8a47-cdd9b33a9dd9 service nova] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.108329] env[61768]: DEBUG oslo_concurrency.lockutils [req-65fd4031-dcd3-44b4-a70e-6805317a2dda req-bc20992d-9c12-4c7f-8a47-cdd9b33a9dd9 service nova] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.108662] env[61768]: DEBUG nova.compute.manager [req-65fd4031-dcd3-44b4-a70e-6805317a2dda req-bc20992d-9c12-4c7f-8a47-cdd9b33a9dd9 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] No waiting events found dispatching network-vif-plugged-bc255400-bfd6-496f-a447-290139231203 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.108939] env[61768]: WARNING nova.compute.manager [req-65fd4031-dcd3-44b4-a70e-6805317a2dda req-bc20992d-9c12-4c7f-8a47-cdd9b33a9dd9 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Received unexpected event network-vif-plugged-bc255400-bfd6-496f-a447-290139231203 for instance with vm_state building and task_state spawning. [ 832.111496] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e8ae137-82c4-4ebe-a5c0-3c3372567727 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.121370] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29e99c49-d1ed-439b-bfb7-16dd6b44b6ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.161742] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b40627-7bb0-4a3f-a53f-762c0e56fdbc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.171386] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228955, 'name': PowerOffVM_Task, 'duration_secs': 0.333911} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.174058] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 832.174343] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance '4f429be7-c594-4a72-aebd-0b746219231c' progress to 17 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 832.179742] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b6e649-e43c-4f97-81ee-0e3cb4b8dda0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.197094] env[61768]: DEBUG nova.compute.provider_tree [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.213793] env[61768]: DEBUG nova.network.neutron [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Successfully updated port: bc255400-bfd6-496f-a447-290139231203 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 832.415712] env[61768]: DEBUG oslo_vmware.api [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228956, 'name': PowerOffVM_Task, 'duration_secs': 0.206289} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.416785] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 832.417121] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 832.417486] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c52ac126-47d0-4071-b1cf-4439c5b622f5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.423761] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228957, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.513050] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 832.513050] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 832.513050] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleting the datastore file [datastore2] 37b729bf-ad5e-4e17-b11b-77b504c049cd {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 832.513394] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36b21edf-4315-46ba-ba66-b77331d5e53c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.522063] env[61768]: DEBUG oslo_vmware.api [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 832.522063] env[61768]: value = "task-1228960" [ 832.522063] env[61768]: _type = "Task" [ 832.522063] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.532670] env[61768]: DEBUG oslo_vmware.api [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.681330] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.681756] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.681960] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.682197] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.682401] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.682602] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.682815] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.683032] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.683325] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.683584] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.683887] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.689946] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a2d8e07-991b-4043-a057-4b95d1326f6a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.712148] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 832.712148] env[61768]: value = "task-1228961" [ 832.712148] env[61768]: _type = "Task" [ 832.712148] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.716105] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-41f6300d-7462-4e11-b32b-8c892e87bafc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.716275] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-41f6300d-7462-4e11-b32b-8c892e87bafc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.716444] env[61768]: DEBUG nova.network.neutron [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.724709] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228961, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.735774] env[61768]: DEBUG nova.scheduler.client.report [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Updated inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 832.736141] env[61768]: DEBUG nova.compute.provider_tree [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 91 to 92 during operation: update_inventory {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 832.736377] env[61768]: DEBUG nova.compute.provider_tree [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.921455] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228957, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.588441} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.921924] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 8e0e13a7-eac5-4176-8536-2906d13c390e/8e0e13a7-eac5-4176-8536-2906d13c390e.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 832.922283] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 832.922655] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2c2b8644-c036-45e9-addd-28b9f95b0213 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.931107] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 832.931107] env[61768]: value = "task-1228962" [ 832.931107] env[61768]: _type = "Task" [ 832.931107] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.941557] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228962, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.037899] env[61768]: DEBUG oslo_vmware.api [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1228960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.275817} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.038930] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 833.039239] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 833.039901] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 833.040203] env[61768]: INFO nova.compute.manager [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Took 1.18 seconds to destroy the instance on the hypervisor. [ 833.040622] env[61768]: DEBUG oslo.service.loopingcall [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.040986] env[61768]: DEBUG nova.compute.manager [-] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.041154] env[61768]: DEBUG nova.network.neutron [-] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 833.225484] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228961, 'name': ReconfigVM_Task, 'duration_secs': 0.223221} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.225739] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance '4f429be7-c594-4a72-aebd-0b746219231c' progress to 33 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 833.241655] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.557s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.242211] env[61768]: DEBUG nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.244830] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.964s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.246345] env[61768]: INFO nova.compute.claims [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.262889] env[61768]: DEBUG nova.network.neutron [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.420046] env[61768]: DEBUG nova.network.neutron [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Updating instance_info_cache with network_info: [{"id": "bc255400-bfd6-496f-a447-290139231203", "address": "fa:16:3e:a2:8c:bf", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc255400-bf", "ovs_interfaceid": "bc255400-bfd6-496f-a447-290139231203", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.441878] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228962, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.1104} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.442172] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 833.442961] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2f6908-fb3d-4376-8332-56d6eab420c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.474608] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 8e0e13a7-eac5-4176-8536-2906d13c390e/8e0e13a7-eac5-4176-8536-2906d13c390e.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 833.475139] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9262d9b3-1e8e-4ba3-9d3c-25e11f773a9e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.506085] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 833.506085] env[61768]: value = "task-1228963" [ 833.506085] env[61768]: _type = "Task" [ 833.506085] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.517631] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228963, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.734102] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.734102] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.734102] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.734560] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.734560] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.735174] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.735174] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.735174] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.735379] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.735417] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.735604] env[61768]: DEBUG nova.virt.hardware [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.741481] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Reconfiguring VM instance instance-00000042 to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 833.741823] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f46e01dd-8c3d-4a20-94e2-76b25ea617ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.756826] env[61768]: DEBUG nova.compute.utils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.760250] env[61768]: DEBUG nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.760436] env[61768]: DEBUG nova.network.neutron [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 833.771626] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 833.771626] env[61768]: value = "task-1228964" [ 833.771626] env[61768]: _type = "Task" [ 833.771626] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.782252] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228964, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.800459] env[61768]: DEBUG nova.network.neutron [-] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.843290] env[61768]: DEBUG nova.policy [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '33db714fc1044225ac0286708a0bb7c0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fa6da53b22194f598ac23dae71b24c07', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.923251] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-41f6300d-7462-4e11-b32b-8c892e87bafc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.923677] env[61768]: DEBUG nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Instance network_info: |[{"id": "bc255400-bfd6-496f-a447-290139231203", "address": "fa:16:3e:a2:8c:bf", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc255400-bf", "ovs_interfaceid": "bc255400-bfd6-496f-a447-290139231203", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.924218] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:8c:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bc255400-bfd6-496f-a447-290139231203', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.932216] env[61768]: DEBUG oslo.service.loopingcall [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.932490] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 833.932760] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ebf5abe2-2d5b-4d45-9710-4904d16a5bd7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.956740] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.956740] env[61768]: value = "task-1228965" [ 833.956740] env[61768]: _type = "Task" [ 833.956740] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.967876] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228965, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.019180] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228963, 'name': ReconfigVM_Task, 'duration_secs': 0.511656} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.019631] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 8e0e13a7-eac5-4176-8536-2906d13c390e/8e0e13a7-eac5-4176-8536-2906d13c390e.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 834.020458] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-19813583-9634-4897-a252-46780147e51e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.028547] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 834.028547] env[61768]: value = "task-1228966" [ 834.028547] env[61768]: _type = "Task" [ 834.028547] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.038731] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228966, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.143748] env[61768]: DEBUG nova.compute.manager [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Received event network-changed-bc255400-bfd6-496f-a447-290139231203 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.144039] env[61768]: DEBUG nova.compute.manager [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Refreshing instance network info cache due to event network-changed-bc255400-bfd6-496f-a447-290139231203. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 834.144310] env[61768]: DEBUG oslo_concurrency.lockutils [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] Acquiring lock "refresh_cache-41f6300d-7462-4e11-b32b-8c892e87bafc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.144528] env[61768]: DEBUG oslo_concurrency.lockutils [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] Acquired lock "refresh_cache-41f6300d-7462-4e11-b32b-8c892e87bafc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.144759] env[61768]: DEBUG nova.network.neutron [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Refreshing network info cache for port bc255400-bfd6-496f-a447-290139231203 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 834.181887] env[61768]: DEBUG nova.network.neutron [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Successfully created port: 7758e5b9-f72b-49f5-84df-7ddcc681d518 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.262092] env[61768]: DEBUG nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.284729] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228964, 'name': ReconfigVM_Task, 'duration_secs': 0.228512} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.285588] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Reconfigured VM instance instance-00000042 to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 834.286787] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41ce1967-fb47-436f-b38b-79d78829d83a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.313087] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 4f429be7-c594-4a72-aebd-0b746219231c/4f429be7-c594-4a72-aebd-0b746219231c.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.317817] env[61768]: INFO nova.compute.manager [-] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Took 1.28 seconds to deallocate network for instance. [ 834.317817] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a6c9360-6973-43be-b178-fad02add4059 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.342282] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 834.342282] env[61768]: value = "task-1228968" [ 834.342282] env[61768]: _type = "Task" [ 834.342282] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.351404] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228968, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.470874] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228965, 'name': CreateVM_Task, 'duration_secs': 0.444074} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.470874] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 834.471910] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.471910] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.471910] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.472210] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2faf290-dfbd-47f5-b922-d6cdf7594e17 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.478492] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 834.478492] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524be5e6-35f1-0026-85e2-d0311b1f9291" [ 834.478492] env[61768]: _type = "Task" [ 834.478492] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.487672] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524be5e6-35f1-0026-85e2-d0311b1f9291, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.539178] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228966, 'name': Rename_Task, 'duration_secs': 0.176655} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.541926] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 834.542409] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5d811dd0-2a23-4a2c-9f86-de0b13bd1593 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.549797] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 834.549797] env[61768]: value = "task-1228969" [ 834.549797] env[61768]: _type = "Task" [ 834.549797] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.562820] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228969, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.620264] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae2fd65-7c78-46ad-a80b-f99d3ac158b4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.628445] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f9c7c49-da38-4326-9daf-19202adb2b69 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.412707] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.430594] env[61768]: DEBUG oslo_vmware.api [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1228969, 'name': PowerOnVM_Task, 'duration_secs': 0.543561} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.430856] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228968, 'name': ReconfigVM_Task, 'duration_secs': 0.285364} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.456711] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 835.456998] env[61768]: INFO nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Took 8.59 seconds to spawn the instance on the hypervisor. [ 835.457210] env[61768]: DEBUG nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 835.457553] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 4f429be7-c594-4a72-aebd-0b746219231c/4f429be7-c594-4a72-aebd-0b746219231c.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 835.457832] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance '4f429be7-c594-4a72-aebd-0b746219231c' progress to 50 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 835.461430] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524be5e6-35f1-0026-85e2-d0311b1f9291, 'name': SearchDatastore_Task, 'duration_secs': 0.01829} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.465015] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b286cf-5cd2-4e76-92b0-858c84e17938 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.468013] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d3f262-6412-421b-9aec-a121a0ef4684 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.470663] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.470901] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.471153] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.471305] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.471487] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.472045] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64992656-a6ff-4c90-8fab-ab6ac9cb8ff0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.482293] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd629026-9cfd-45d3-9e3f-2cf0e2adb14d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.490131] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.490336] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.491622] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a2c5ac1-1d35-4572-920d-4d508b65515f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.501823] env[61768]: DEBUG nova.compute.provider_tree [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.507879] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 835.507879] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52697549-348a-b57f-108d-50ceb83209f1" [ 835.507879] env[61768]: _type = "Task" [ 835.507879] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.519091] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52697549-348a-b57f-108d-50ceb83209f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.646851] env[61768]: DEBUG nova.compute.manager [req-846d6f91-354e-473a-a626-2babdd571c6f req-d082d11c-36b1-433f-83d1-e5cffc27b212 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Received event network-vif-plugged-7758e5b9-f72b-49f5-84df-7ddcc681d518 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.647055] env[61768]: DEBUG oslo_concurrency.lockutils [req-846d6f91-354e-473a-a626-2babdd571c6f req-d082d11c-36b1-433f-83d1-e5cffc27b212 service nova] Acquiring lock "aebe41ad-496d-4a53-b023-c8df2bca04f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.647279] env[61768]: DEBUG oslo_concurrency.lockutils [req-846d6f91-354e-473a-a626-2babdd571c6f req-d082d11c-36b1-433f-83d1-e5cffc27b212 service nova] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.647456] env[61768]: DEBUG oslo_concurrency.lockutils [req-846d6f91-354e-473a-a626-2babdd571c6f req-d082d11c-36b1-433f-83d1-e5cffc27b212 service nova] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.647633] env[61768]: DEBUG nova.compute.manager [req-846d6f91-354e-473a-a626-2babdd571c6f req-d082d11c-36b1-433f-83d1-e5cffc27b212 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] No waiting events found dispatching network-vif-plugged-7758e5b9-f72b-49f5-84df-7ddcc681d518 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.647800] env[61768]: WARNING nova.compute.manager [req-846d6f91-354e-473a-a626-2babdd571c6f req-d082d11c-36b1-433f-83d1-e5cffc27b212 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Received unexpected event network-vif-plugged-7758e5b9-f72b-49f5-84df-7ddcc681d518 for instance with vm_state building and task_state spawning. [ 835.676096] env[61768]: DEBUG nova.network.neutron [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Updated VIF entry in instance network info cache for port bc255400-bfd6-496f-a447-290139231203. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 835.676505] env[61768]: DEBUG nova.network.neutron [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Updating instance_info_cache with network_info: [{"id": "bc255400-bfd6-496f-a447-290139231203", "address": "fa:16:3e:a2:8c:bf", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbc255400-bf", "ovs_interfaceid": "bc255400-bfd6-496f-a447-290139231203", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.737588] env[61768]: DEBUG nova.network.neutron [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Successfully updated port: 7758e5b9-f72b-49f5-84df-7ddcc681d518 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.915387] env[61768]: DEBUG nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.942995] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.943313] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.943484] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.943683] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.943846] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.944020] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.944255] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.944435] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.944766] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.945114] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.945383] env[61768]: DEBUG nova.virt.hardware [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.946512] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e00a44-590d-460f-b2df-961009770a61 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.955209] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841895c5-bd29-487b-b11a-de873e49b311 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.970876] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62b47c9e-b7a6-48b8-9b2c-b3cd9427d3de {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.990830] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9b67ac-94c2-4435-ac97-f4afaff977a0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.001462] env[61768]: INFO nova.compute.manager [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Took 38.42 seconds to build instance. [ 836.017561] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance '4f429be7-c594-4a72-aebd-0b746219231c' progress to 67 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 836.031399] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52697549-348a-b57f-108d-50ceb83209f1, 'name': SearchDatastore_Task, 'duration_secs': 0.015681} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.032814] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8273ebe1-a10d-492b-b321-181922d3076d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.039805] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 836.039805] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]521bdac6-e235-d2ca-c133-bf7193a21c93" [ 836.039805] env[61768]: _type = "Task" [ 836.039805] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.049181] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]521bdac6-e235-d2ca-c133-bf7193a21c93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.056427] env[61768]: DEBUG nova.scheduler.client.report [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updated inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with generation 92 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 836.056704] env[61768]: DEBUG nova.compute.provider_tree [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 92 to 93 during operation: update_inventory {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 836.056893] env[61768]: DEBUG nova.compute.provider_tree [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.179480] env[61768]: DEBUG oslo_concurrency.lockutils [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] Releasing lock "refresh_cache-41f6300d-7462-4e11-b32b-8c892e87bafc" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.179968] env[61768]: DEBUG nova.compute.manager [req-c7f65286-c751-4aee-94f6-c3eecb590e64 req-13e6d352-b29c-4bcf-b4ba-3787b2a77b18 service nova] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Received event network-vif-deleted-43cc41b6-c86c-4fce-a565-3585c3014f7e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.240991] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.241261] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.241489] env[61768]: DEBUG nova.network.neutron [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.521931] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e441ee3f-f2fe-4d73-a0c6-6f49d1d3f27e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.679s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.554573] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]521bdac6-e235-d2ca-c133-bf7193a21c93, 'name': SearchDatastore_Task, 'duration_secs': 0.014028} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.554573] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.554573] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 41f6300d-7462-4e11-b32b-8c892e87bafc/41f6300d-7462-4e11-b32b-8c892e87bafc.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 836.554573] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f83abb70-bb6a-4e95-8200-8c7c806902c6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.561261] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 836.561261] env[61768]: value = "task-1228971" [ 836.561261] env[61768]: _type = "Task" [ 836.561261] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.565137] env[61768]: DEBUG nova.network.neutron [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Port 0e23d603-ab56-4c07-b562-335a39497bac binding to destination host cpu-1 is already ACTIVE {{(pid=61768) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 836.568285] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.322s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.568285] env[61768]: DEBUG nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.571681] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.983s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.571681] env[61768]: DEBUG nova.objects.instance [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lazy-loading 'resources' on Instance uuid a36b542c-334f-46b1-9030-ee5c9f8c55c1 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 836.579952] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228971, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.784157] env[61768]: DEBUG nova.network.neutron [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.912878] env[61768]: DEBUG nova.compute.manager [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Received event network-changed-e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 836.913310] env[61768]: DEBUG nova.compute.manager [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Refreshing instance network info cache due to event network-changed-e02bb27b-f94f-4f93-859a-5af2646daaaa. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 836.913393] env[61768]: DEBUG oslo_concurrency.lockutils [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] Acquiring lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.913483] env[61768]: DEBUG oslo_concurrency.lockutils [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] Acquired lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.913662] env[61768]: DEBUG nova.network.neutron [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Refreshing network info cache for port e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.017401] env[61768]: DEBUG nova.network.neutron [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Updating instance_info_cache with network_info: [{"id": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "address": "fa:16:3e:99:c1:46", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7758e5b9-f7", "ovs_interfaceid": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.078395] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228971, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.079813] env[61768]: DEBUG nova.compute.utils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 837.085292] env[61768]: DEBUG nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 837.085751] env[61768]: DEBUG nova.network.neutron [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 837.165693] env[61768]: DEBUG nova.policy [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7f7389d526c4ce9bc8ecf062b9921e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '180da0bcd5954eb2bc89a20cf8f42bb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 837.419960] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ae8acf-ea26-4e8e-9ae9-1026a364d98f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.429568] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8ba3cc-bb57-459a-9f93-476cee7b5a7e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.464185] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-624c377d-a861-4139-81b3-d0400f302cbf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.474020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4e7c80-ad42-452d-b45d-e02e498c57ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.487997] env[61768]: DEBUG nova.compute.provider_tree [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.520558] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.520975] env[61768]: DEBUG nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Instance network_info: |[{"id": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "address": "fa:16:3e:99:c1:46", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7758e5b9-f7", "ovs_interfaceid": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.521483] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:99:c1:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7758e5b9-f72b-49f5-84df-7ddcc681d518', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.529308] env[61768]: DEBUG oslo.service.loopingcall [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.529563] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.529744] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f7dd813f-85a2-4e3b-9166-00e4e465a483 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.545596] env[61768]: DEBUG nova.network.neutron [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Successfully created port: a8acff1f-7c19-4525-bc2d-3d6345758f75 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.553911] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.553911] env[61768]: value = "task-1228972" [ 837.553911] env[61768]: _type = "Task" [ 837.553911] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.563603] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228972, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.573377] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228971, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.566242} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.573683] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 41f6300d-7462-4e11-b32b-8c892e87bafc/41f6300d-7462-4e11-b32b-8c892e87bafc.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 837.573937] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.574262] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a723540f-fe96-48e9-9f84-4c9f26dbe0ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.582738] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 837.582738] env[61768]: value = "task-1228973" [ 837.582738] env[61768]: _type = "Task" [ 837.582738] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.594563] env[61768]: DEBUG nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.611684] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "4f429be7-c594-4a72-aebd-0b746219231c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.612012] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.612299] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.616015] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228973, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.680990] env[61768]: DEBUG nova.compute.manager [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Received event network-changed-7758e5b9-f72b-49f5-84df-7ddcc681d518 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.681122] env[61768]: DEBUG nova.compute.manager [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Refreshing instance network info cache due to event network-changed-7758e5b9-f72b-49f5-84df-7ddcc681d518. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.681364] env[61768]: DEBUG oslo_concurrency.lockutils [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] Acquiring lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.681568] env[61768]: DEBUG oslo_concurrency.lockutils [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] Acquired lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.681756] env[61768]: DEBUG nova.network.neutron [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Refreshing network info cache for port 7758e5b9-f72b-49f5-84df-7ddcc681d518 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.766811] env[61768]: DEBUG nova.network.neutron [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Updated VIF entry in instance network info cache for port e02bb27b-f94f-4f93-859a-5af2646daaaa. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 837.766811] env[61768]: DEBUG nova.network.neutron [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Updating instance_info_cache with network_info: [{"id": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "address": "fa:16:3e:39:0c:b3", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.145", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape02bb27b-f9", "ovs_interfaceid": "e02bb27b-f94f-4f93-859a-5af2646daaaa", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.991808] env[61768]: DEBUG nova.scheduler.client.report [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.067040] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228972, 'name': CreateVM_Task, 'duration_secs': 0.440283} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.067202] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.067809] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.067995] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.068356] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.068665] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fee75b43-3cf5-411d-9706-dd341aaf050d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.074615] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 838.074615] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52afb97b-e4d3-4ebc-6a8d-2f2456a7e7b9" [ 838.074615] env[61768]: _type = "Task" [ 838.074615] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.084396] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52afb97b-e4d3-4ebc-6a8d-2f2456a7e7b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.092843] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228973, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087574} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.093135] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.093976] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2af7a70-d8f4-468d-b451-5cebf3480676 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.117260] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] 41f6300d-7462-4e11-b32b-8c892e87bafc/41f6300d-7462-4e11-b32b-8c892e87bafc.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.122564] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0505abb-ecc1-4a71-9761-6a2e841fc503 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.151319] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 838.151319] env[61768]: value = "task-1228974" [ 838.151319] env[61768]: _type = "Task" [ 838.151319] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.161534] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228974, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.209727] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.209968] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.210254] env[61768]: DEBUG nova.network.neutron [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 838.268674] env[61768]: DEBUG oslo_concurrency.lockutils [req-296b84d1-6872-440a-911b-d73f40e4b93b req-e8c3354f-a6f5-4117-9467-f73cf6186a97 service nova] Releasing lock "refresh_cache-8e0e13a7-eac5-4176-8536-2906d13c390e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.497917] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.927s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.500805] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.664s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.502249] env[61768]: DEBUG nova.objects.instance [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lazy-loading 'resources' on Instance uuid d928f14c-b94a-4abf-a053-51015f3bc6c5 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.533245] env[61768]: INFO nova.scheduler.client.report [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleted allocations for instance a36b542c-334f-46b1-9030-ee5c9f8c55c1 [ 838.568952] env[61768]: DEBUG nova.network.neutron [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Updated VIF entry in instance network info cache for port 7758e5b9-f72b-49f5-84df-7ddcc681d518. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.569355] env[61768]: DEBUG nova.network.neutron [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Updating instance_info_cache with network_info: [{"id": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "address": "fa:16:3e:99:c1:46", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7758e5b9-f7", "ovs_interfaceid": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.589836] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52afb97b-e4d3-4ebc-6a8d-2f2456a7e7b9, 'name': SearchDatastore_Task, 'duration_secs': 0.014507} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.590174] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.590420] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.590679] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.590845] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.591048] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.591327] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6d0145f-8140-4cab-ae48-637480088a32 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.603335] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.603578] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 838.604664] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c934b76-50c0-4b0a-8dfd-9f68c3158628 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.611922] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 838.611922] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524b8be5-e524-7806-0f03-96737f315757" [ 838.611922] env[61768]: _type = "Task" [ 838.611922] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.621304] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524b8be5-e524-7806-0f03-96737f315757, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.641253] env[61768]: DEBUG nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.670659] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228974, 'name': ReconfigVM_Task, 'duration_secs': 0.419662} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.671251] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Reconfigured VM instance instance-0000004e to attach disk [datastore1] 41f6300d-7462-4e11-b32b-8c892e87bafc/41f6300d-7462-4e11-b32b-8c892e87bafc.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 838.673812] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6757c4f818faa8f38482586df8d3f816',container_format='bare',created_at=2024-10-09T19:57:15Z,direct_url=,disk_format='vmdk',id=6c2ff2ef-4663-4449-a0eb-4863e7b27db2,min_disk=1,min_ram=0,name='tempest-test-snap-1584606764',owner='180da0bcd5954eb2bc89a20cf8f42bb5',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-09T19:57:31Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.674061] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.674231] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.674420] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.674571] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.674725] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.674973] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.675172] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.675351] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.675520] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.675698] env[61768]: DEBUG nova.virt.hardware [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.675970] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98886b3b-b9bb-4c77-ac84-e2fe275ee52a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.678225] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2f82c3c-332b-4f61-b39b-1dd3501cd08f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.688581] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e535a96-5e3f-44d9-9338-6bc939be4225 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.692559] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 838.692559] env[61768]: value = "task-1228976" [ 838.692559] env[61768]: _type = "Task" [ 838.692559] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.709890] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228976, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.847610] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.848015] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.958645] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52510471-ea17-2ca9-5eab-bf09932be0a9/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 838.959731] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffda4c41-994d-4a88-a34e-bc1c1826a875 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.966389] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52510471-ea17-2ca9-5eab-bf09932be0a9/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 838.966681] env[61768]: ERROR oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52510471-ea17-2ca9-5eab-bf09932be0a9/disk-0.vmdk due to incomplete transfer. [ 838.966811] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-85468e3f-7d6b-47ef-b633-296eddd34bf3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.974371] env[61768]: DEBUG oslo_vmware.rw_handles [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52510471-ea17-2ca9-5eab-bf09932be0a9/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 838.974599] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Uploaded image 1e83b86d-089f-4b79-940c-ef76d051d4f3 to the Glance image server {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 838.977929] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 838.978219] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-05b4c2ee-4348-4f39-8367-6e42778b8fec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.993416] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 838.993416] env[61768]: value = "task-1228977" [ 838.993416] env[61768]: _type = "Task" [ 838.993416] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.002688] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228977, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.047503] env[61768]: DEBUG oslo_concurrency.lockutils [None req-abb71437-46a5-4741-8add-76e0956557ac tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "a36b542c-334f-46b1-9030-ee5c9f8c55c1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.592s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.072860] env[61768]: DEBUG oslo_concurrency.lockutils [req-19c9fcab-6480-4cb3-974b-ff6d478061fd req-1cc3c8c3-c10f-4b54-a0a0-0706059e6523 service nova] Releasing lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.124159] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524b8be5-e524-7806-0f03-96737f315757, 'name': SearchDatastore_Task, 'duration_secs': 0.015817} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.124537] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4d3aad8-bdf3-4878-b2f8-ae3697c99c94 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.130288] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 839.130288] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52349ef0-921b-53ae-f2e7-16e725218bc3" [ 839.130288] env[61768]: _type = "Task" [ 839.130288] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.139818] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52349ef0-921b-53ae-f2e7-16e725218bc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.143025] env[61768]: DEBUG nova.network.neutron [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance_info_cache with network_info: [{"id": "0e23d603-ab56-4c07-b562-335a39497bac", "address": "fa:16:3e:3d:a9:43", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e23d603-ab", "ovs_interfaceid": "0e23d603-ab56-4c07-b562-335a39497bac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.176742] env[61768]: DEBUG nova.network.neutron [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Successfully updated port: a8acff1f-7c19-4525-bc2d-3d6345758f75 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.206450] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228976, 'name': Rename_Task, 'duration_secs': 0.197541} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.209344] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 839.210041] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9e3c761a-aa80-49e0-991f-baf0f3932908 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.216731] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 839.216731] env[61768]: value = "task-1228978" [ 839.216731] env[61768]: _type = "Task" [ 839.216731] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.228263] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228978, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.288695] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cf9d20d-4f8e-4a6c-ada0-2a7bdf16cbf9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.297144] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c98943-d6ca-4a88-a87b-e117084d0baf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.329797] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44adbde9-d34b-41d0-96c1-af4f5cec4aa5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.338403] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d78dac7-8784-4392-8e5d-2550cff22d57 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.352683] env[61768]: DEBUG nova.compute.utils [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.355664] env[61768]: DEBUG nova.compute.provider_tree [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.504030] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228977, 'name': Destroy_Task, 'duration_secs': 0.389888} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.504321] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Destroyed the VM [ 839.504647] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 839.504912] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-06a0bafd-805e-4e88-9268-e2b27c22ef59 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.513088] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 839.513088] env[61768]: value = "task-1228979" [ 839.513088] env[61768]: _type = "Task" [ 839.513088] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.522181] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228979, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.641919] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52349ef0-921b-53ae-f2e7-16e725218bc3, 'name': SearchDatastore_Task, 'duration_secs': 0.013188} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.642293] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.642387] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] aebe41ad-496d-4a53-b023-c8df2bca04f8/aebe41ad-496d-4a53-b023-c8df2bca04f8.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.642664] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d289b245-564b-4260-aea5-46d106c31ef3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.646291] env[61768]: DEBUG oslo_concurrency.lockutils [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.651965] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 839.651965] env[61768]: value = "task-1228980" [ 839.651965] env[61768]: _type = "Task" [ 839.651965] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.661537] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228980, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.678152] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "refresh_cache-86e24eb9-6a55-4658-9e62-64713b0289d7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.678317] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "refresh_cache-86e24eb9-6a55-4658-9e62-64713b0289d7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.678543] env[61768]: DEBUG nova.network.neutron [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 839.730672] env[61768]: DEBUG oslo_vmware.api [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1228978, 'name': PowerOnVM_Task, 'duration_secs': 0.502159} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.731086] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 839.731361] env[61768]: INFO nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Took 8.02 seconds to spawn the instance on the hypervisor. [ 839.731851] env[61768]: DEBUG nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 839.732706] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72fbf924-4b3f-460d-b040-566eb3fbf3d7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.856243] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.858299] env[61768]: DEBUG nova.scheduler.client.report [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.008443] env[61768]: DEBUG nova.compute.manager [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Received event network-vif-plugged-a8acff1f-7c19-4525-bc2d-3d6345758f75 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.008839] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] Acquiring lock "86e24eb9-6a55-4658-9e62-64713b0289d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.009161] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.009432] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.009737] env[61768]: DEBUG nova.compute.manager [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] No waiting events found dispatching network-vif-plugged-a8acff1f-7c19-4525-bc2d-3d6345758f75 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.010070] env[61768]: WARNING nova.compute.manager [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Received unexpected event network-vif-plugged-a8acff1f-7c19-4525-bc2d-3d6345758f75 for instance with vm_state building and task_state spawning. [ 840.010628] env[61768]: DEBUG nova.compute.manager [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Received event network-changed-a8acff1f-7c19-4525-bc2d-3d6345758f75 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.010628] env[61768]: DEBUG nova.compute.manager [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Refreshing instance network info cache due to event network-changed-a8acff1f-7c19-4525-bc2d-3d6345758f75. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.010859] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] Acquiring lock "refresh_cache-86e24eb9-6a55-4658-9e62-64713b0289d7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.030380] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228979, 'name': RemoveSnapshot_Task, 'duration_secs': 0.462697} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.030717] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 840.031193] env[61768]: DEBUG nova.compute.manager [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.032468] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71cd8d0a-1c75-4e36-bb35-7ed42b5b5c0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.168781] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228980, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.179346] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e011579-54c8-444f-a9eb-f62e8e82dc68 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.203011] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba53a24-3d78-488b-8be4-029bb5bb9031 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.211416] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance '4f429be7-c594-4a72-aebd-0b746219231c' progress to 83 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 840.217777] env[61768]: DEBUG nova.network.neutron [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.258053] env[61768]: INFO nova.compute.manager [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Took 38.85 seconds to build instance. [ 840.366138] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.863s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.367325] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.352s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.368967] env[61768]: INFO nova.compute.claims [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.389288] env[61768]: INFO nova.scheduler.client.report [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleted allocations for instance d928f14c-b94a-4abf-a053-51015f3bc6c5 [ 840.417534] env[61768]: DEBUG nova.network.neutron [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Updating instance_info_cache with network_info: [{"id": "a8acff1f-7c19-4525-bc2d-3d6345758f75", "address": "fa:16:3e:54:af:d5", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8acff1f-7c", "ovs_interfaceid": "a8acff1f-7c19-4525-bc2d-3d6345758f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.551858] env[61768]: INFO nova.compute.manager [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Shelve offloading [ 840.553566] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 840.553822] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e41bf75-5e6b-4945-a89e-b70d46038ec9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.561922] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 840.561922] env[61768]: value = "task-1228981" [ 840.561922] env[61768]: _type = "Task" [ 840.561922] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.572063] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228981, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.666428] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228980, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.586033} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.666757] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] aebe41ad-496d-4a53-b023-c8df2bca04f8/aebe41ad-496d-4a53-b023-c8df2bca04f8.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.667018] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.667293] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1c326d27-c29d-458d-9847-28009433a1e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.674717] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 840.674717] env[61768]: value = "task-1228982" [ 840.674717] env[61768]: _type = "Task" [ 840.674717] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.686969] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228982, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.718024] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.718149] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c694bf69-3586-49a2-a697-fe1d5a5a29ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.725976] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 840.725976] env[61768]: value = "task-1228983" [ 840.725976] env[61768]: _type = "Task" [ 840.725976] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.735184] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.761091] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a142672b-c160-4848-bc65-f767156ff183 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.686s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.897942] env[61768]: DEBUG oslo_concurrency.lockutils [None req-565c8fca-730f-4c26-a1d5-7903be07673f tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "d928f14c-b94a-4abf-a053-51015f3bc6c5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.031s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.919023] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "refresh_cache-86e24eb9-6a55-4658-9e62-64713b0289d7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.919280] env[61768]: DEBUG nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Instance network_info: |[{"id": "a8acff1f-7c19-4525-bc2d-3d6345758f75", "address": "fa:16:3e:54:af:d5", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8acff1f-7c", "ovs_interfaceid": "a8acff1f-7c19-4525-bc2d-3d6345758f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 840.919612] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] Acquired lock "refresh_cache-86e24eb9-6a55-4658-9e62-64713b0289d7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.919843] env[61768]: DEBUG nova.network.neutron [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Refreshing network info cache for port a8acff1f-7c19-4525-bc2d-3d6345758f75 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 840.920979] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:af:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a8acff1f-7c19-4525-bc2d-3d6345758f75', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 840.929775] env[61768]: DEBUG oslo.service.loopingcall [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.930837] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 840.931138] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2afe9a9-78bc-4f99-a099-59e1e2e5e568 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.946706] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.947207] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.947449] env[61768]: INFO nova.compute.manager [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Attaching volume a38ffb49-6325-4242-819b-ba4b116389d5 to /dev/sdb [ 840.958101] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 840.958101] env[61768]: value = "task-1228984" [ 840.958101] env[61768]: _type = "Task" [ 840.958101] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.966448] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228984, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.982384] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6ffa78-fe88-48d8-8d13-234dae71e417 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.993834] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421f0c42-9e48-4775-b89c-242624d75975 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.012405] env[61768]: DEBUG nova.virt.block_device [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating existing volume attachment record: e0fe5127-b2d1-4dd9-ab5f-3117c83f5839 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 841.074100] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 841.074709] env[61768]: DEBUG nova.compute.manager [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.075207] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bedffa4-c65b-4280-9489-377ae37d0ccc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.082943] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.083422] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.083422] env[61768]: DEBUG nova.network.neutron [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 841.098065] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "d13673f2-e99f-400b-bcdc-b7c38e197610" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.098394] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.185962] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228982, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066702} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.186225] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.187124] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ff641c-7a4b-4ed7-b22b-1870def2e6d0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.213592] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] aebe41ad-496d-4a53-b023-c8df2bca04f8/aebe41ad-496d-4a53-b023-c8df2bca04f8.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.214040] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52c1bf5d-38a6-459a-a9ef-b2c380751fc2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.239256] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228983, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.249480] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 841.249480] env[61768]: value = "task-1228986" [ 841.249480] env[61768]: _type = "Task" [ 841.249480] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.258455] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228986, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.443440] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "41f6300d-7462-4e11-b32b-8c892e87bafc" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.443828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.467798] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228984, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.604938] env[61768]: DEBUG nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.664015] env[61768]: DEBUG nova.network.neutron [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Updated VIF entry in instance network info cache for port a8acff1f-7c19-4525-bc2d-3d6345758f75. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 841.664454] env[61768]: DEBUG nova.network.neutron [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Updating instance_info_cache with network_info: [{"id": "a8acff1f-7c19-4525-bc2d-3d6345758f75", "address": "fa:16:3e:54:af:d5", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa8acff1f-7c", "ovs_interfaceid": "a8acff1f-7c19-4525-bc2d-3d6345758f75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.676181] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1e35ef-ff7d-4da6-b18b-a95fcca85111 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.685266] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfe7e5f-d581-44c0-9251-6bf5c4cc53b4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.721307] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30fe1ee8-e2da-443c-8daf-ed530762ab98 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.734105] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2065f3-7a25-434a-9982-9b57d03894ed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.744275] env[61768]: DEBUG oslo_vmware.api [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1228983, 'name': PowerOnVM_Task, 'duration_secs': 0.73755} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.752030] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.753068] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-16941bf6-a622-422a-911d-f9c169586b43 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance '4f429be7-c594-4a72-aebd-0b746219231c' progress to 100 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 841.757284] env[61768]: DEBUG nova.compute.provider_tree [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.772582] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228986, 'name': ReconfigVM_Task, 'duration_secs': 0.369212} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.772874] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Reconfigured VM instance instance-0000004f to attach disk [datastore1] aebe41ad-496d-4a53-b023-c8df2bca04f8/aebe41ad-496d-4a53-b023-c8df2bca04f8.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.773542] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d8037e3-ca3e-4cb4-b152-f8e069d740ca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.781116] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 841.781116] env[61768]: value = "task-1228989" [ 841.781116] env[61768]: _type = "Task" [ 841.781116] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.790855] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228989, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.904470] env[61768]: DEBUG nova.network.neutron [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.947452] env[61768]: DEBUG nova.compute.utils [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 841.970290] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1228984, 'name': CreateVM_Task, 'duration_secs': 0.620127} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.970489] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 841.971204] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.971370] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.971764] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 841.972031] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cbcbb84-f9a7-471c-8ade-990895554310 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.976844] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 841.976844] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d5c575-2b51-ecc0-95d2-6058cd314b1a" [ 841.976844] env[61768]: _type = "Task" [ 841.976844] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.984943] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d5c575-2b51-ecc0-95d2-6058cd314b1a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.131293] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.168423] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbe22260-7ce3-4ac6-8893-2109250970a3 req-018997d8-7ba1-43a4-92d2-cc2e875df2df service nova] Releasing lock "refresh_cache-86e24eb9-6a55-4658-9e62-64713b0289d7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.194600] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquiring lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.195340] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.265516] env[61768]: DEBUG nova.scheduler.client.report [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.292256] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228989, 'name': Rename_Task, 'duration_secs': 0.156754} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.292588] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 842.292897] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7e5591f6-aa4d-4aaf-a608-2f04fc516c77 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.300886] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 842.300886] env[61768]: value = "task-1228990" [ 842.300886] env[61768]: _type = "Task" [ 842.300886] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.310078] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228990, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.407826] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.450431] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.492687] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.493049] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Processing image 6c2ff2ef-4663-4449-a0eb-4863e7b27db2 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 842.493247] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2/6c2ff2ef-4663-4449-a0eb-4863e7b27db2.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.493389] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2/6c2ff2ef-4663-4449-a0eb-4863e7b27db2.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.493611] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 842.493858] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-965e1664-e3e1-45cf-ba98-52b428d18f34 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.504344] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 842.504548] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 842.505748] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9ede3a30-2fe1-4f45-b497-dc38276bc895 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.512082] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 842.512082] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fc0df3-884b-bbf9-d0bd-f3235c008681" [ 842.512082] env[61768]: _type = "Task" [ 842.512082] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.522162] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fc0df3-884b-bbf9-d0bd-f3235c008681, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.697053] env[61768]: DEBUG nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 842.770882] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.771669] env[61768]: DEBUG nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.782968] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 14.932s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.783217] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.004s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.787020] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 842.787020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.028s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.787020] env[61768]: DEBUG nova.objects.instance [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lazy-loading 'resources' on Instance uuid d4a765bd-a56d-44b7-8db3-c081832d58c9 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.787020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b842c8-b209-475b-bc8d-a02438f47e44 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.796874] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41dfc1f8-6e7c-4adc-b607-8e227adcffa2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.822146] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228990, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.823121] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4af6c7-44b9-4aec-a302-1f45266032d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.831320] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fe75dcf-2de2-4a61-a73d-55e64642bfa3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.840683] env[61768]: DEBUG nova.compute.manager [req-d5015e1e-1641-4b58-8fd2-5c3484027346 req-e867f6e3-5c9b-4c65-a1e9-1ff137a483d4 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-vif-unplugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.840999] env[61768]: DEBUG oslo_concurrency.lockutils [req-d5015e1e-1641-4b58-8fd2-5c3484027346 req-e867f6e3-5c9b-4c65-a1e9-1ff137a483d4 service nova] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.841184] env[61768]: DEBUG oslo_concurrency.lockutils [req-d5015e1e-1641-4b58-8fd2-5c3484027346 req-e867f6e3-5c9b-4c65-a1e9-1ff137a483d4 service nova] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.841356] env[61768]: DEBUG oslo_concurrency.lockutils [req-d5015e1e-1641-4b58-8fd2-5c3484027346 req-e867f6e3-5c9b-4c65-a1e9-1ff137a483d4 service nova] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.841540] env[61768]: DEBUG nova.compute.manager [req-d5015e1e-1641-4b58-8fd2-5c3484027346 req-e867f6e3-5c9b-4c65-a1e9-1ff137a483d4 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] No waiting events found dispatching network-vif-unplugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.841722] env[61768]: WARNING nova.compute.manager [req-d5015e1e-1641-4b58-8fd2-5c3484027346 req-e867f6e3-5c9b-4c65-a1e9-1ff137a483d4 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received unexpected event network-vif-unplugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 for instance with vm_state shelved and task_state shelving_offloading. [ 842.870408] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180253MB free_disk=150GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 842.870595] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.871920] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 842.873193] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fb73c4-afc7-4c76-81ee-a163513a71d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.881490] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 842.881767] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c42f4b84-942a-4a64-9913-c2333eeb6d90 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.980230] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "b6e4f743-5692-4b46-892c-6c8917ccef98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.980589] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.980880] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "b6e4f743-5692-4b46-892c-6c8917ccef98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.981275] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.981520] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.983966] env[61768]: INFO nova.compute.manager [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Terminating instance [ 842.986496] env[61768]: DEBUG nova.compute.manager [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 842.986787] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 842.987837] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b46fe328-40a4-4302-b9bc-08c5945757f8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.996630] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 842.996887] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31309260-0f80-4581-9320-11b1dbb941e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.004871] env[61768]: DEBUG oslo_vmware.api [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 843.004871] env[61768]: value = "task-1228992" [ 843.004871] env[61768]: _type = "Task" [ 843.004871] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.020743] env[61768]: DEBUG oslo_vmware.api [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228992, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.032593] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Preparing fetch location {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 843.033314] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Fetch image to [datastore1] OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310/OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310.vmdk {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 843.033314] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Downloading stream optimized image 6c2ff2ef-4663-4449-a0eb-4863e7b27db2 to [datastore1] OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310/OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310.vmdk on the data store datastore1 as vApp {{(pid=61768) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 843.033314] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Downloading image file data 6c2ff2ef-4663-4449-a0eb-4863e7b27db2 to the ESX as VM named 'OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310' {{(pid=61768) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 843.055340] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "356a69b0-7898-440d-9473-7b4572ed2315" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.055826] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "356a69b0-7898-440d-9473-7b4572ed2315" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.056513] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "356a69b0-7898-440d-9473-7b4572ed2315-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.056513] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "356a69b0-7898-440d-9473-7b4572ed2315-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.056680] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "356a69b0-7898-440d-9473-7b4572ed2315-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.059921] env[61768]: INFO nova.compute.manager [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Terminating instance [ 843.062055] env[61768]: DEBUG nova.compute.manager [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.062417] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 843.063634] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173a258d-cd7a-4ebf-9af5-7c4d2bd9faf8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.086221] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 843.086589] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 843.086898] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleting the datastore file [datastore1] 62cddff5-d499-4ef3-869d-3cdb0328640d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.088357] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-828bde8a-54b2-4a51-8d42-9c156d54cbfc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.094602] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 843.094924] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e0ff795-37df-41e7-ae18-6ecaa75548f3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.097771] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 843.097771] env[61768]: value = "task-1228993" [ 843.097771] env[61768]: _type = "Task" [ 843.097771] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.108919] env[61768]: DEBUG oslo_vmware.api [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 843.108919] env[61768]: value = "task-1228994" [ 843.108919] env[61768]: _type = "Task" [ 843.108919] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.109206] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228993, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.120358] env[61768]: DEBUG oslo_vmware.api [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.132224] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 843.132224] env[61768]: value = "resgroup-9" [ 843.132224] env[61768]: _type = "ResourcePool" [ 843.132224] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 843.132800] env[61768]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9c75299f-4daf-4444-9799-c5f1f36f94ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.156782] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lease: (returnval){ [ 843.156782] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52631459-28bc-0a75-11b7-c737ecf43a7b" [ 843.156782] env[61768]: _type = "HttpNfcLease" [ 843.156782] env[61768]: } obtained for vApp import into resource pool (val){ [ 843.156782] env[61768]: value = "resgroup-9" [ 843.156782] env[61768]: _type = "ResourcePool" [ 843.156782] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 843.157404] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the lease: (returnval){ [ 843.157404] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52631459-28bc-0a75-11b7-c737ecf43a7b" [ 843.157404] env[61768]: _type = "HttpNfcLease" [ 843.157404] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 843.166343] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.166343] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52631459-28bc-0a75-11b7-c737ecf43a7b" [ 843.166343] env[61768]: _type = "HttpNfcLease" [ 843.166343] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 843.225767] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.287627] env[61768]: DEBUG nova.compute.utils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.291539] env[61768]: DEBUG nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.291539] env[61768]: DEBUG nova.network.neutron [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 843.321065] env[61768]: DEBUG oslo_vmware.api [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1228990, 'name': PowerOnVM_Task, 'duration_secs': 0.558772} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.321760] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 843.322027] env[61768]: INFO nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Took 7.41 seconds to spawn the instance on the hypervisor. [ 843.322247] env[61768]: DEBUG nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.323089] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19908ff0-a67a-4909-82d2-ed30c64c98aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.354485] env[61768]: DEBUG nova.policy [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.519911] env[61768]: DEBUG oslo_vmware.api [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228992, 'name': PowerOffVM_Task, 'duration_secs': 0.249565} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.520238] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.520429] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 843.520695] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f14448f6-7297-4a9d-ad64-6200f64c45d2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.522733] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "41f6300d-7462-4e11-b32b-8c892e87bafc" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.523031] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.523228] env[61768]: INFO nova.compute.manager [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Attaching volume 318a620b-f812-4fc3-8024-611696fac705 to /dev/sdb [ 843.564369] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a65a8e91-94dc-4515-b638-70b67fe7db8d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.574422] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31934495-b494-4f93-93b3-83363ae186be {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.592839] env[61768]: DEBUG nova.virt.block_device [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Updating existing volume attachment record: 7674e688-a667-40ef-abe1-05e0b9845db2 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 843.609557] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 843.610873] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 843.610873] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleting the datastore file [datastore2] b6e4f743-5692-4b46-892c-6c8917ccef98 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.611066] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6dc81720-439a-451e-940c-8485332ff2e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.619951] env[61768]: DEBUG oslo_vmware.api [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1228993, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231773} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.623949] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.624308] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 843.624598] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 843.636844] env[61768]: DEBUG oslo_vmware.api [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 843.636844] env[61768]: value = "task-1228998" [ 843.636844] env[61768]: _type = "Task" [ 843.636844] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.639440] env[61768]: DEBUG oslo_vmware.api [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228994, 'name': PowerOffVM_Task, 'duration_secs': 0.233313} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.640991] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.641281] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 843.643271] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee62ccc-2d0c-446e-92a9-74671a022bf3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.650728] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3245f3d8-e097-4254-8b55-bd39246e2dc7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.653790] env[61768]: INFO nova.scheduler.client.report [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted allocations for instance 62cddff5-d499-4ef3-869d-3cdb0328640d [ 843.666106] env[61768]: DEBUG oslo_vmware.api [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228998, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.667506] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb691604-c4de-49fd-9054-861e32326bb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.672667] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 843.672667] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52631459-28bc-0a75-11b7-c737ecf43a7b" [ 843.672667] env[61768]: _type = "HttpNfcLease" [ 843.672667] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 843.706249] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f2fa8b-72b1-43fe-bc5e-6b7129a6b779 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.717028] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f58bab00-5b98-4b76-9c52-fa3540689649 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.734036] env[61768]: DEBUG nova.compute.provider_tree [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.740634] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 843.740634] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 843.740634] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleting the datastore file [datastore2] 356a69b0-7898-440d-9473-7b4572ed2315 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.740980] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ef3a0e73-1d40-42c0-b9f0-b56ca133c520 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.749431] env[61768]: DEBUG oslo_vmware.api [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for the task: (returnval){ [ 843.749431] env[61768]: value = "task-1229000" [ 843.749431] env[61768]: _type = "Task" [ 843.749431] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.760857] env[61768]: DEBUG oslo_vmware.api [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1229000, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.797978] env[61768]: DEBUG nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.816795] env[61768]: DEBUG nova.network.neutron [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Successfully created port: b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.850907] env[61768]: INFO nova.compute.manager [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Took 36.02 seconds to build instance. [ 844.101520] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "4f429be7-c594-4a72-aebd-0b746219231c" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.101520] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.102413] env[61768]: DEBUG nova.compute.manager [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Going to confirm migration 1 {{(pid=61768) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 844.153635] env[61768]: DEBUG oslo_vmware.api [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1228998, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197493} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.153971] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.154128] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 844.154743] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 844.154743] env[61768]: INFO nova.compute.manager [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Took 1.17 seconds to destroy the instance on the hypervisor. [ 844.154743] env[61768]: DEBUG oslo.service.loopingcall [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.155426] env[61768]: DEBUG nova.compute.manager [-] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.155585] env[61768]: DEBUG nova.network.neutron [-] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 844.158107] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.166935] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 844.166935] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52631459-28bc-0a75-11b7-c737ecf43a7b" [ 844.166935] env[61768]: _type = "HttpNfcLease" [ 844.166935] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 844.166935] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 844.166935] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52631459-28bc-0a75-11b7-c737ecf43a7b" [ 844.166935] env[61768]: _type = "HttpNfcLease" [ 844.166935] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 844.167823] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02309974-c0f5-4a80-b0b9-18f03dccaa81 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.176029] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52581ad2-9a8d-cc50-0dc4-6458773e07dc/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 844.176228] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52581ad2-9a8d-cc50-0dc4-6458773e07dc/disk-0.vmdk. {{(pid=61768) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 844.239413] env[61768]: DEBUG nova.scheduler.client.report [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.249377] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-fce79fc8-2b49-4a40-be75-a10800881d5d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.262986] env[61768]: DEBUG oslo_vmware.api [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Task: {'id': task-1229000, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.299739} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.264099] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.264340] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 844.264532] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 844.264710] env[61768]: INFO nova.compute.manager [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Took 1.20 seconds to destroy the instance on the hypervisor. [ 844.265095] env[61768]: DEBUG oslo.service.loopingcall [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.265777] env[61768]: DEBUG nova.compute.manager [-] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.265777] env[61768]: DEBUG nova.network.neutron [-] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 844.353301] env[61768]: DEBUG oslo_concurrency.lockutils [None req-fe648f31-13d9-4246-90a0-3d19fc912382 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.533s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.474924] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] Acquiring lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.475212] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] Acquired lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.475482] env[61768]: DEBUG nova.network.neutron [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.611124] env[61768]: DEBUG nova.compute.manager [req-c9d4a64d-47f7-4a72-a212-de645d5dced1 req-4a971231-c17b-48e2-961a-0bbcd7da16a8 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Received event network-vif-deleted-9cef3fe6-e733-4c02-82f1-2d914170b440 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.611802] env[61768]: INFO nova.compute.manager [req-c9d4a64d-47f7-4a72-a212-de645d5dced1 req-4a971231-c17b-48e2-961a-0bbcd7da16a8 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Neutron deleted interface 9cef3fe6-e733-4c02-82f1-2d914170b440; detaching it from the instance and deleting it from the info cache [ 844.611802] env[61768]: DEBUG nova.network.neutron [req-c9d4a64d-47f7-4a72-a212-de645d5dced1 req-4a971231-c17b-48e2-961a-0bbcd7da16a8 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.665072] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.665378] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.665828] env[61768]: DEBUG nova.network.neutron [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.665828] env[61768]: DEBUG nova.objects.instance [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lazy-loading 'info_cache' on Instance uuid 4f429be7-c594-4a72-aebd-0b746219231c {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.744844] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.961s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.747615] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.954s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.747863] env[61768]: DEBUG nova.objects.instance [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lazy-loading 'resources' on Instance uuid 3aa722e5-0818-40ae-8220-223d920a7386 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 844.773736] env[61768]: INFO nova.scheduler.client.report [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted allocations for instance d4a765bd-a56d-44b7-8db3-c081832d58c9 [ 844.809743] env[61768]: DEBUG nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.839515] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.839873] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.840055] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.840329] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.840570] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.840709] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.840970] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.841195] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.841374] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.841571] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.841737] env[61768]: DEBUG nova.virt.hardware [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.843401] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c7baa2-6043-4e4d-ab28-2e69ea3a22ba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.858426] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55fb07c-f39c-4365-a11f-3dcac49926b0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.947401] env[61768]: DEBUG nova.network.neutron [-] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.033331] env[61768]: DEBUG nova.network.neutron [-] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.114360] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-227dc44c-c967-4e60-a83b-afec306fa9b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.128393] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d771a027-cf69-4612-8793-643e0a3ac4df {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.162353] env[61768]: DEBUG nova.compute.manager [req-c9d4a64d-47f7-4a72-a212-de645d5dced1 req-4a971231-c17b-48e2-961a-0bbcd7da16a8 service nova] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Detach interface failed, port_id=9cef3fe6-e733-4c02-82f1-2d914170b440, reason: Instance b6e4f743-5692-4b46-892c-6c8917ccef98 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 845.287126] env[61768]: DEBUG oslo_concurrency.lockutils [None req-733195f8-8312-4d70-9217-1fa3eecbfa6a tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "d4a765bd-a56d-44b7-8db3-c081832d58c9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.680s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.451216] env[61768]: INFO nova.compute.manager [-] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Took 1.30 seconds to deallocate network for instance. [ 845.463343] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Completed reading data from the image iterator. {{(pid=61768) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 845.463728] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52581ad2-9a8d-cc50-0dc4-6458773e07dc/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 845.465047] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6332a8-b82f-40d9-a1cd-ad9b4023072e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.477405] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52581ad2-9a8d-cc50-0dc4-6458773e07dc/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 845.477645] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52581ad2-9a8d-cc50-0dc4-6458773e07dc/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 845.477896] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-dc2a36f0-33c7-4a7a-90cd-95921890349f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.540105] env[61768]: INFO nova.compute.manager [-] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Took 1.27 seconds to deallocate network for instance. [ 845.577333] env[61768]: DEBUG nova.network.neutron [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Updating instance_info_cache with network_info: [{"id": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "address": "fa:16:3e:99:c1:46", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7758e5b9-f7", "ovs_interfaceid": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.580254] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 845.580729] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265474', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'name': 'volume-a38ffb49-6325-4242-819b-ba4b116389d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7697ecda-ef30-447d-a92a-3f5cb3cc9118', 'attached_at': '', 'detached_at': '', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'serial': 'a38ffb49-6325-4242-819b-ba4b116389d5'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 845.582634] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c65f9c7-d2b6-4b8e-983a-ff9bdef94fab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.605648] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5735ebeb-f6fd-4374-9edb-895e4a1739cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.635891] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] volume-a38ffb49-6325-4242-819b-ba4b116389d5/volume-a38ffb49-6325-4242-819b-ba4b116389d5.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.642022] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dba3ab3-bebf-4666-ae07-277f8abd84ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.660783] env[61768]: DEBUG nova.compute.manager [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 845.660783] env[61768]: DEBUG nova.compute.manager [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing instance network info cache due to event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 845.660783] env[61768]: DEBUG oslo_concurrency.lockutils [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.661044] env[61768]: DEBUG oslo_concurrency.lockutils [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.661195] env[61768]: DEBUG nova.network.neutron [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 845.675562] env[61768]: DEBUG oslo_vmware.api [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 845.675562] env[61768]: value = "task-1229004" [ 845.675562] env[61768]: _type = "Task" [ 845.675562] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.681676] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25efb7a-f4ca-4074-83af-7a9a78164362 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.691340] env[61768]: DEBUG oslo_vmware.api [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229004, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.695647] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3c7914-a635-41f8-8053-05aee071baf5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.699698] env[61768]: DEBUG oslo_vmware.rw_handles [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52581ad2-9a8d-cc50-0dc4-6458773e07dc/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 845.700011] env[61768]: INFO nova.virt.vmwareapi.images [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Downloaded image file data 6c2ff2ef-4663-4449-a0eb-4863e7b27db2 [ 845.701445] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ddbd37c-a01b-47f9-9e7c-594d2998dd40 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.750535] env[61768]: DEBUG nova.network.neutron [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Successfully updated port: b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 845.752292] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de5a4319-d468-44ad-b428-45e05c802e18 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.754760] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f68c9f76-58ec-447e-a133-80c727eade8b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.764433] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08614c50-4ed2-4153-9a4f-a89a38d9d13d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.780814] env[61768]: DEBUG nova.compute.provider_tree [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 845.787449] env[61768]: INFO nova.virt.vmwareapi.images [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] The imported VM was unregistered [ 845.790056] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Caching image {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 845.790311] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating directory with path [datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.791229] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb19a020-d48a-4711-852a-48ad257e0843 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.804011] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Created directory with path [datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.804230] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310/OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310.vmdk to [datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2/6c2ff2ef-4663-4449-a0eb-4863e7b27db2.vmdk. {{(pid=61768) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 845.804483] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-38d003c2-6e76-4fbf-9e6a-21841626413e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.813129] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 845.813129] env[61768]: value = "task-1229006" [ 845.813129] env[61768]: _type = "Task" [ 845.813129] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.822719] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229006, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.970709] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.015370] env[61768]: DEBUG nova.network.neutron [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance_info_cache with network_info: [{"id": "0e23d603-ab56-4c07-b562-335a39497bac", "address": "fa:16:3e:3d:a9:43", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0e23d603-ab", "ovs_interfaceid": "0e23d603-ab56-4c07-b562-335a39497bac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.055242] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.083345] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] Releasing lock "refresh_cache-aebe41ad-496d-4a53-b023-c8df2bca04f8" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.083345] env[61768]: DEBUG nova.compute.manager [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Inject network info {{(pid=61768) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 846.083345] env[61768]: DEBUG nova.compute.manager [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] network_info to inject: |[{"id": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "address": "fa:16:3e:99:c1:46", "network": {"id": "8ac77af9-ceb4-4fe7-af6a-e9385e592c1b", "bridge": "br-int", "label": "tempest-ServersAdminTestJSON-1783588209-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fa6da53b22194f598ac23dae71b24c07", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ce058b2d-df85-481c-a996-cc179d534f1a", "external-id": "nsx-vlan-transportzone-671", "segmentation_id": 671, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7758e5b9-f7", "ovs_interfaceid": "7758e5b9-f72b-49f5-84df-7ddcc681d518", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 846.089673] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Reconfiguring VM instance to set the machine id {{(pid=61768) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 846.089673] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fe54636-d39a-4653-912e-c3ae5c5590d0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.111914] env[61768]: DEBUG oslo_vmware.api [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] Waiting for the task: (returnval){ [ 846.111914] env[61768]: value = "task-1229008" [ 846.111914] env[61768]: _type = "Task" [ 846.111914] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.127053] env[61768]: DEBUG oslo_vmware.api [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] Task: {'id': task-1229008, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.189970] env[61768]: DEBUG oslo_vmware.api [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.257822] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.258465] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.258465] env[61768]: DEBUG nova.network.neutron [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 846.284433] env[61768]: DEBUG nova.scheduler.client.report [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.330131] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229006, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.475170] env[61768]: DEBUG nova.network.neutron [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updated VIF entry in instance network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 846.475637] env[61768]: DEBUG nova.network.neutron [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.519102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-4f429be7-c594-4a72-aebd-0b746219231c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.519400] env[61768]: DEBUG nova.objects.instance [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lazy-loading 'migration_context' on Instance uuid 4f429be7-c594-4a72-aebd-0b746219231c {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.630728] env[61768]: DEBUG oslo_vmware.api [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] Task: {'id': task-1229008, 'name': ReconfigVM_Task, 'duration_secs': 0.480472} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.631262] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8f385edd-59dc-40cb-aa1a-fb99ae406d0d tempest-ServersAdminTestJSON-198141517 tempest-ServersAdminTestJSON-198141517-project-admin] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Reconfigured VM instance to set the machine id {{(pid=61768) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 846.688747] env[61768]: DEBUG oslo_vmware.api [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229004, 'name': ReconfigVM_Task, 'duration_secs': 1.001329} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.689109] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Reconfigured VM instance instance-00000043 to attach disk [datastore2] volume-a38ffb49-6325-4242-819b-ba4b116389d5/volume-a38ffb49-6325-4242-819b-ba4b116389d5.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.694021] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adcb1402-07d9-40dd-a9cf-65aacb21788e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.716015] env[61768]: DEBUG oslo_vmware.api [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 846.716015] env[61768]: value = "task-1229009" [ 846.716015] env[61768]: _type = "Task" [ 846.716015] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.731101] env[61768]: DEBUG oslo_vmware.api [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229009, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.793054] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.045s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.796289] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.383s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.796456] env[61768]: DEBUG nova.objects.instance [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'resources' on Instance uuid 37b729bf-ad5e-4e17-b11b-77b504c049cd {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.799685] env[61768]: DEBUG nova.network.neutron [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.814231] env[61768]: INFO nova.scheduler.client.report [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Deleted allocations for instance 3aa722e5-0818-40ae-8220-223d920a7386 [ 846.837087] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229006, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.981032] env[61768]: DEBUG oslo_concurrency.lockutils [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.981032] env[61768]: DEBUG nova.compute.manager [req-4ff56e67-e623-4105-92eb-004706023c0c req-5382444a-a119-4b1b-80bb-00c08e77d3c5 service nova] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Received event network-vif-deleted-97d08d00-af43-4a90-b759-76b75806aa51 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.984630] env[61768]: DEBUG nova.network.neutron [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.023379] env[61768]: DEBUG nova.objects.base [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Object Instance<4f429be7-c594-4a72-aebd-0b746219231c> lazy-loaded attributes: info_cache,migration_context {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 847.024893] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5af6dd2c-6775-4a8a-b87d-74a0dc9a4d7b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.050086] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-563c890a-b8f1-4e27-80c4-a3ca5ef71220 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.060222] env[61768]: DEBUG oslo_vmware.api [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 847.060222] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a3589-48ac-c43b-b867-7be8c5d41bc9" [ 847.060222] env[61768]: _type = "Task" [ 847.060222] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.073165] env[61768]: DEBUG oslo_vmware.api [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a3589-48ac-c43b-b867-7be8c5d41bc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.232317] env[61768]: DEBUG oslo_vmware.api [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229009, 'name': ReconfigVM_Task, 'duration_secs': 0.466551} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.232660] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265474', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'name': 'volume-a38ffb49-6325-4242-819b-ba4b116389d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7697ecda-ef30-447d-a92a-3f5cb3cc9118', 'attached_at': '', 'detached_at': '', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'serial': 'a38ffb49-6325-4242-819b-ba4b116389d5'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 847.338118] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229006, 'name': MoveVirtualDisk_Task} progress is 60%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.340794] env[61768]: DEBUG oslo_concurrency.lockutils [None req-312d6b65-e32b-4c37-ab78-10da9e5209be tempest-MultipleCreateTestJSON-1611722169 tempest-MultipleCreateTestJSON-1611722169-project-member] Lock "3aa722e5-0818-40ae-8220-223d920a7386" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.661s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.378811] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.487269] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.487768] env[61768]: DEBUG nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Instance network_info: |[{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 847.488400] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:99:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4ad17b4-d077-472c-be99-bd42736950e4', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 847.499255] env[61768]: DEBUG oslo.service.loopingcall [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.502693] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 847.503260] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-70a0e0f6-46a7-4da4-a4b7-db8a8eeddf88 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.532048] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 847.532048] env[61768]: value = "task-1229010" [ 847.532048] env[61768]: _type = "Task" [ 847.532048] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.547484] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229010, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.576179] env[61768]: DEBUG oslo_vmware.api [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a3589-48ac-c43b-b867-7be8c5d41bc9, 'name': SearchDatastore_Task, 'duration_secs': 0.013395} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.579422] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.619351] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05910b78-507f-43e0-8a72-c4602ab9518f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.628273] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7acc8053-c605-408c-ae70-e75ddc49addc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.664746] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278a1735-81b5-4a7c-88cc-aaef7ccc0b9b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.675519] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbce89be-8801-4c4f-94c9-e98d2a6fed09 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.695937] env[61768]: DEBUG nova.compute.provider_tree [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 847.750338] env[61768]: DEBUG nova.compute.manager [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-vif-plugged-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.751065] env[61768]: DEBUG oslo_concurrency.lockutils [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] Acquiring lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.751331] env[61768]: DEBUG oslo_concurrency.lockutils [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.751520] env[61768]: DEBUG oslo_concurrency.lockutils [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.751761] env[61768]: DEBUG nova.compute.manager [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] No waiting events found dispatching network-vif-plugged-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.751882] env[61768]: WARNING nova.compute.manager [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received unexpected event network-vif-plugged-b4ad17b4-d077-472c-be99-bd42736950e4 for instance with vm_state building and task_state spawning. [ 847.752063] env[61768]: DEBUG nova.compute.manager [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.752229] env[61768]: DEBUG nova.compute.manager [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing instance network info cache due to event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.752423] env[61768]: DEBUG oslo_concurrency.lockutils [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.752567] env[61768]: DEBUG oslo_concurrency.lockutils [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.752737] env[61768]: DEBUG nova.network.neutron [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 847.830358] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229006, 'name': MoveVirtualDisk_Task} progress is 80%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.851046] env[61768]: INFO nova.compute.manager [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Rebuilding instance [ 847.910242] env[61768]: DEBUG nova.compute.manager [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 847.911455] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2aa873a-68aa-4985-8971-751c47e230ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.051568] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229010, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.164027] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 848.164027] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265477', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'name': 'volume-318a620b-f812-4fc3-8024-611696fac705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41f6300d-7462-4e11-b32b-8c892e87bafc', 'attached_at': '', 'detached_at': '', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'serial': '318a620b-f812-4fc3-8024-611696fac705'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 848.164027] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe59d863-abbc-484d-b482-b9e183134608 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.188571] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647a7664-a6c7-488c-a08a-1de90b70c7ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.210653] env[61768]: DEBUG nova.scheduler.client.report [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 848.222477] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] volume-318a620b-f812-4fc3-8024-611696fac705/volume-318a620b-f812-4fc3-8024-611696fac705.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.223225] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6f50d40-2938-4b4a-94d7-f40979a338a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.243407] env[61768]: DEBUG oslo_vmware.api [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 848.243407] env[61768]: value = "task-1229011" [ 848.243407] env[61768]: _type = "Task" [ 848.243407] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.254150] env[61768]: DEBUG oslo_vmware.api [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229011, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.287205] env[61768]: DEBUG nova.objects.instance [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 848.328566] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229006, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.429971] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 848.433117] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f143496c-39dd-48cf-b3e5-d5e560f700ee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.442965] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 848.442965] env[61768]: value = "task-1229012" [ 848.442965] env[61768]: _type = "Task" [ 848.442965] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.453074] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.509506] env[61768]: DEBUG nova.network.neutron [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updated VIF entry in instance network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 848.510305] env[61768]: DEBUG nova.network.neutron [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.548092] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229010, 'name': CreateVM_Task, 'duration_secs': 0.631809} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.548092] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 848.548092] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.548092] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.548092] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 848.548092] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26f94bd1-e295-4938-807d-9ca339e3112a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.554351] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 848.554351] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52552947-e1c0-95fb-cd0b-5e5215bd5cfb" [ 848.554351] env[61768]: _type = "Task" [ 848.554351] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.565128] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52552947-e1c0-95fb-cd0b-5e5215bd5cfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.727136] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.729693] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.599s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.735177] env[61768]: INFO nova.compute.claims [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.755231] env[61768]: DEBUG oslo_vmware.api [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229011, 'name': ReconfigVM_Task, 'duration_secs': 0.377548} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.755757] env[61768]: INFO nova.scheduler.client.report [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted allocations for instance 37b729bf-ad5e-4e17-b11b-77b504c049cd [ 848.760017] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Reconfigured VM instance instance-0000004e to attach disk [datastore2] volume-318a620b-f812-4fc3-8024-611696fac705/volume-318a620b-f812-4fc3-8024-611696fac705.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.763994] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb54fa03-fc1b-4169-977c-c5dcdffbcf50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.782094] env[61768]: DEBUG oslo_vmware.api [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 848.782094] env[61768]: value = "task-1229013" [ 848.782094] env[61768]: _type = "Task" [ 848.782094] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.797990] env[61768]: DEBUG oslo_vmware.api [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229013, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.797990] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6ba5e495-a59e-4551-b962-ebeb47f542a6 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.850s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.836626] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229006, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.687169} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.837299] env[61768]: INFO nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310/OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310.vmdk to [datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2/6c2ff2ef-4663-4449-a0eb-4863e7b27db2.vmdk. [ 848.837652] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Cleaning up location [datastore1] OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 848.837949] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_f5317b1b-1509-44d4-a9c8-83b21068b310 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 848.838573] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b404410b-887a-4949-9423-9a428199328b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.847768] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 848.847768] env[61768]: value = "task-1229014" [ 848.847768] env[61768]: _type = "Task" [ 848.847768] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.863183] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.959657] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229012, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.013348] env[61768]: DEBUG oslo_concurrency.lockutils [req-97d1f9b2-03a9-4e73-8f7d-854a79272b0f req-3c94da22-6154-4c20-8df4-26e182c9a7a4 service nova] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.067405] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52552947-e1c0-95fb-cd0b-5e5215bd5cfb, 'name': SearchDatastore_Task, 'duration_secs': 0.038869} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.067740] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.068068] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 849.068346] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.068517] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.068910] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 849.069282] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19ca8ac5-6a6e-48ba-aecd-f82bc44a7242 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.080319] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 849.080506] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 849.081770] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20c4d57c-5515-4054-a10d-989eef0ffbce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.087759] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 849.087759] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b28e87-81cd-7cdb-5886-aa9492d34279" [ 849.087759] env[61768]: _type = "Task" [ 849.087759] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.107014] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b28e87-81cd-7cdb-5886-aa9492d34279, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.279992] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f1e821ef-5a4a-4e66-a500-c70fcb6cfcf0 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "37b729bf-ad5e-4e17-b11b-77b504c049cd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.424s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.298701] env[61768]: DEBUG oslo_vmware.api [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229013, 'name': ReconfigVM_Task, 'duration_secs': 0.163014} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.299711] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265477', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'name': 'volume-318a620b-f812-4fc3-8024-611696fac705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41f6300d-7462-4e11-b32b-8c892e87bafc', 'attached_at': '', 'detached_at': '', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'serial': '318a620b-f812-4fc3-8024-611696fac705'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 849.358610] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.109295} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.359077] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.359452] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2/6c2ff2ef-4663-4449-a0eb-4863e7b27db2.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.359824] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2/6c2ff2ef-4663-4449-a0eb-4863e7b27db2.vmdk to [datastore1] 86e24eb9-6a55-4658-9e62-64713b0289d7/86e24eb9-6a55-4658-9e62-64713b0289d7.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 849.360249] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7b461e2b-3064-46dd-b251-a213e5f09c82 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.372020] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 849.372020] env[61768]: value = "task-1229015" [ 849.372020] env[61768]: _type = "Task" [ 849.372020] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.380623] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229015, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.390459] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.390969] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.391333] env[61768]: DEBUG nova.compute.manager [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.392314] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c22a77-e57f-4d40-9637-56683ddc8b4c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.399345] env[61768]: DEBUG nova.compute.manager [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61768) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 849.400246] env[61768]: DEBUG nova.objects.instance [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.455861] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229012, 'name': PowerOffVM_Task, 'duration_secs': 0.558236} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.456225] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 849.457010] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 849.458050] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8181c92-72be-49e3-afcd-4620a149f13b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.466332] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 849.466627] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0e0b045d-b1e4-4a90-946b-f9a5fe0f0986 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.546616] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 849.547078] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 849.547441] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleting the datastore file [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 849.547781] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-050376cb-5504-4821-a7b2-dbefc4c1160e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.556679] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 849.556679] env[61768]: value = "task-1229017" [ 849.556679] env[61768]: _type = "Task" [ 849.556679] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.564482] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229017, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.600205] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b28e87-81cd-7cdb-5886-aa9492d34279, 'name': SearchDatastore_Task, 'duration_secs': 0.02308} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.601395] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb93b537-099d-4944-946b-9dd3d469d1fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.607642] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 849.607642] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522af6e5-ae7a-ec9b-ae69-ab0d0d7542e0" [ 849.607642] env[61768]: _type = "Task" [ 849.607642] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.616322] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522af6e5-ae7a-ec9b-ae69-ab0d0d7542e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.884470] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229015, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.905621] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 849.906082] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70a2de27-6979-4efe-930e-b0762453b2e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.916645] env[61768]: DEBUG oslo_vmware.api [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 849.916645] env[61768]: value = "task-1229018" [ 849.916645] env[61768]: _type = "Task" [ 849.916645] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.932984] env[61768]: DEBUG oslo_vmware.api [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.016298] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ac2683-4bf9-4996-94ce-15b2137687f8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.025552] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74899065-ec8f-4000-8042-aa0d6779d99f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.067870] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3830be73-6c6d-45ad-bd0d-e1369c4009a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.080985] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229017, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.381524} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.082297] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34757e27-2b99-4b59-bd1d-609cdb921aaf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.086403] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 850.086632] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 850.086881] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 850.106825] env[61768]: DEBUG nova.compute.provider_tree [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.126297] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522af6e5-ae7a-ec9b-ae69-ab0d0d7542e0, 'name': SearchDatastore_Task, 'duration_secs': 0.042417} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.126656] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.127039] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] ecd08348-b22f-49c5-b13f-b187506e38d6/ecd08348-b22f-49c5-b13f-b187506e38d6.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 850.127419] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-75794034-0d43-49eb-8533-f9e67686ab82 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.141868] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 850.141868] env[61768]: value = "task-1229019" [ 850.141868] env[61768]: _type = "Task" [ 850.141868] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.158614] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229019, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.243935] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "12f92761-45b6-49c6-96bd-1c0fa05521c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.243935] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.344389] env[61768]: DEBUG nova.objects.instance [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'flavor' on Instance uuid 41f6300d-7462-4e11-b32b-8c892e87bafc {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 850.384846] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229015, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.435070] env[61768]: DEBUG oslo_vmware.api [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.649439] env[61768]: ERROR nova.scheduler.client.report [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [req-437a212d-45c6-4977-99a6-cdfdfcea2e5e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-437a212d-45c6-4977-99a6-cdfdfcea2e5e"}]} [ 850.655732] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229019, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.668566] env[61768]: DEBUG nova.scheduler.client.report [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 850.686788] env[61768]: DEBUG nova.scheduler.client.report [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 850.687141] env[61768]: DEBUG nova.compute.provider_tree [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.703413] env[61768]: DEBUG nova.scheduler.client.report [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 850.728456] env[61768]: DEBUG nova.scheduler.client.report [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 850.746022] env[61768]: DEBUG nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 850.852344] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aeb93206-77c7-4b12-8e98-c5c2df536b2a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.329s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.887561] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229015, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.935700] env[61768]: DEBUG oslo_vmware.api [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.044386] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc94849-282c-4f5e-9f55-a38f67f7f5f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.056409] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d310c7-cdf3-45c8-b1d6-d29a2efc61ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.102585] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c48012d6-a952-42bc-b04f-4b2998fb4df8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.114079] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c5d4fa-5d68-4bba-ab0c-3334b15faf6c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.139681] env[61768]: DEBUG nova.compute.provider_tree [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.156118] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.156326] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.156717] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.156840] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.157171] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.157421] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.157718] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.157940] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.158244] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.158485] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.158801] env[61768]: DEBUG nova.virt.hardware [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.160302] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-201306f4-35c8-472c-be68-0f2e6261904b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.169496] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229019, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.820935} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.170423] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] ecd08348-b22f-49c5-b13f-b187506e38d6/ecd08348-b22f-49c5-b13f-b187506e38d6.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 851.170737] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 851.171054] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87eadd79-b635-4a1c-b2de-40a40e9c51d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.181751] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ff82463-5a7b-4405-8fff-e596a41cb3a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.187549] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 851.187549] env[61768]: value = "task-1229020" [ 851.187549] env[61768]: _type = "Task" [ 851.187549] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.201881] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:c4:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2db4501-9ba5-4717-8ce1-2565bb20c603', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 851.211416] env[61768]: DEBUG oslo.service.loopingcall [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 851.215920] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 851.216335] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229020, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.216599] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d41c2c6-f3de-4d2e-acf1-da00dee9a42c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.242211] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 851.242211] env[61768]: value = "task-1229021" [ 851.242211] env[61768]: _type = "Task" [ 851.242211] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.256297] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229021, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.271752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.385615] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229015, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.434869] env[61768]: DEBUG oslo_vmware.api [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229018, 'name': PowerOffVM_Task, 'duration_secs': 1.215794} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.435116] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 851.435320] env[61768]: DEBUG nova.compute.manager [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 851.436398] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3089306-6b73-4d2e-a35c-4716cb35ec52 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.462799] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "41f6300d-7462-4e11-b32b-8c892e87bafc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.462799] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.463031] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "41f6300d-7462-4e11-b32b-8c892e87bafc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.463230] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.465198] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.466124] env[61768]: INFO nova.compute.manager [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Terminating instance [ 851.468724] env[61768]: DEBUG nova.compute.manager [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 851.468977] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 851.469353] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-462c0c1c-b00b-4ffd-86e9-b7ca8c610ab4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.479974] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 851.479974] env[61768]: value = "task-1229022" [ 851.479974] env[61768]: _type = "Task" [ 851.479974] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.492322] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229022, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.643785] env[61768]: DEBUG nova.scheduler.client.report [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.705698] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229020, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082677} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.705698] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 851.705698] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2e59ff-4de5-48f7-8231-a6b42592ad3d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.733287] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] ecd08348-b22f-49c5-b13f-b187506e38d6/ecd08348-b22f-49c5-b13f-b187506e38d6.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 851.734150] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b70a7162-e09c-4654-bc76-4012cee772c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.764027] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229021, 'name': CreateVM_Task, 'duration_secs': 0.40291} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.764027] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 851.764027] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 851.764027] env[61768]: value = "task-1229023" [ 851.764027] env[61768]: _type = "Task" [ 851.764027] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.764899] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.765148] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.765550] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 851.765947] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5db3597f-6a01-4550-af04-f18df4b5cdb6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.781873] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229023, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.782414] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 851.782414] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c27cc7-1350-9666-fe89-403503456479" [ 851.782414] env[61768]: _type = "Task" [ 851.782414] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.794791] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c27cc7-1350-9666-fe89-403503456479, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.886846] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229015, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.952781] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4733518e-18d6-4850-a75b-f13168ca1b8c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.562s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.996263] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229022, 'name': PowerOffVM_Task, 'duration_secs': 0.271012} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.997149] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 851.997149] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 851.997149] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265477', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'name': 'volume-318a620b-f812-4fc3-8024-611696fac705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41f6300d-7462-4e11-b32b-8c892e87bafc', 'attached_at': '', 'detached_at': '', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'serial': '318a620b-f812-4fc3-8024-611696fac705'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 851.998011] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b8cb30-923b-4d76-bb91-aba2a13bbfb1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.024058] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10c6ccc-d2c9-4fb8-b7c0-dc3ece2315da {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.034742] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2594d7-8056-4f33-bce8-fde49edde9e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.060402] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614e0e0d-c066-47be-8957-8798bf2a13f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.080472] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] The volume has not been displaced from its original location: [datastore2] volume-318a620b-f812-4fc3-8024-611696fac705/volume-318a620b-f812-4fc3-8024-611696fac705.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 852.086079] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 852.086465] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-821171de-5529-4bd8-b6ca-3e54e69af56c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.106409] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 852.106409] env[61768]: value = "task-1229024" [ 852.106409] env[61768]: _type = "Task" [ 852.106409] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.115086] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229024, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.150893] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.421s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.151525] env[61768]: DEBUG nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.154260] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 9.284s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.277175] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229023, 'name': ReconfigVM_Task, 'duration_secs': 0.412973} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.277498] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Reconfigured VM instance instance-00000051 to attach disk [datastore2] ecd08348-b22f-49c5-b13f-b187506e38d6/ecd08348-b22f-49c5-b13f-b187506e38d6.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.278172] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b7fb2bd-75a5-426e-bd92-24fdb3d4c492 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.287611] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 852.287611] env[61768]: value = "task-1229025" [ 852.287611] env[61768]: _type = "Task" [ 852.287611] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.294497] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c27cc7-1350-9666-fe89-403503456479, 'name': SearchDatastore_Task, 'duration_secs': 0.016747} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.295685] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.295685] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 852.295685] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.295885] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.296033] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 852.296308] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3896b1dd-aa35-45de-a918-c191dfa97ac6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.301438] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229025, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.309044] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 852.309044] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 852.309793] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df4776db-0db1-4674-a574-dc90bc52ae2e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.315666] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 852.315666] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526cf51f-2a3c-bf9c-ffba-f3100ca99057" [ 852.315666] env[61768]: _type = "Task" [ 852.315666] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.323888] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526cf51f-2a3c-bf9c-ffba-f3100ca99057, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.384809] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229015, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.70248} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.385109] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/6c2ff2ef-4663-4449-a0eb-4863e7b27db2/6c2ff2ef-4663-4449-a0eb-4863e7b27db2.vmdk to [datastore1] 86e24eb9-6a55-4658-9e62-64713b0289d7/86e24eb9-6a55-4658-9e62-64713b0289d7.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 852.385882] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df038e6-69c5-43d5-99ee-2cd34bbd3cfe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.408185] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Reconfiguring VM instance instance-00000050 to attach disk [datastore1] 86e24eb9-6a55-4658-9e62-64713b0289d7/86e24eb9-6a55-4658-9e62-64713b0289d7.vmdk or device None with type streamOptimized {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.408503] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff5bde52-986f-492e-b37e-3442f7ffebc2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.429018] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 852.429018] env[61768]: value = "task-1229026" [ 852.429018] env[61768]: _type = "Task" [ 852.429018] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.438440] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229026, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.617526] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229024, 'name': ReconfigVM_Task, 'duration_secs': 0.206199} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.617821] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 852.623180] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1db29ed6-2762-44cd-82c2-4b64fc706fd8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.641696] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 852.641696] env[61768]: value = "task-1229027" [ 852.641696] env[61768]: _type = "Task" [ 852.641696] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.649455] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229027, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.657195] env[61768]: DEBUG nova.compute.utils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.668172] env[61768]: DEBUG nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.668361] env[61768]: DEBUG nova.network.neutron [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 852.725444] env[61768]: DEBUG nova.policy [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '62522bd3731a47bc928a0c4038c447d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '06ce9d5af7e740ff882439e4486a0aad', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.801900] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229025, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.826691] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526cf51f-2a3c-bf9c-ffba-f3100ca99057, 'name': SearchDatastore_Task, 'duration_secs': 0.01255} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.827531] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92488961-b383-4568-949b-14ed1c5bb6cb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.833415] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 852.833415] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52555244-0778-de15-7023-dbd292da3b83" [ 852.833415] env[61768]: _type = "Task" [ 852.833415] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.841974] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52555244-0778-de15-7023-dbd292da3b83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.939352] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229026, 'name': ReconfigVM_Task, 'duration_secs': 0.297962} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.939659] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Reconfigured VM instance instance-00000050 to attach disk [datastore1] 86e24eb9-6a55-4658-9e62-64713b0289d7/86e24eb9-6a55-4658-9e62-64713b0289d7.vmdk or device None with type streamOptimized {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 852.941207] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b941d2e5-290a-4d85-9989-4b3f0e175bde {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.949687] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 852.949687] env[61768]: value = "task-1229028" [ 852.949687] env[61768]: _type = "Task" [ 852.949687] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.963177] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229028, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.035124] env[61768]: DEBUG nova.network.neutron [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Successfully created port: be0213d6-1e3b-41cf-a6c5-b37ac80034f3 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 853.091671] env[61768]: DEBUG nova.objects.instance [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.155485] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229027, 'name': ReconfigVM_Task, 'duration_secs': 0.456283} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.155895] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265477', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'name': 'volume-318a620b-f812-4fc3-8024-611696fac705', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '41f6300d-7462-4e11-b32b-8c892e87bafc', 'attached_at': '', 'detached_at': '', 'volume_id': '318a620b-f812-4fc3-8024-611696fac705', 'serial': '318a620b-f812-4fc3-8024-611696fac705'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 853.156315] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 853.157411] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb2d63c-0ede-4a4d-ab3a-e7057509589e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.166234] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 853.166479] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-09b4087b-a593-418b-8a64-46dd0fe8cffa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.171615] env[61768]: DEBUG nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.176662] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Applying migration context for instance 4f429be7-c594-4a72-aebd-0b746219231c as it has an incoming, in-progress migration 05095ce1-1d9e-4bdc-8f8d-2b88b84020e3. Migration status is confirming {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 853.176662] env[61768]: INFO nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating resource usage from migration 05095ce1-1d9e-4bdc-8f8d-2b88b84020e3 [ 853.200568] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 2c85de5d-0111-40a6-a8bc-69c7eba0393d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 3d96e47d-fc09-439e-bf0e-e60fc87ae958 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7697ecda-ef30-447d-a92a-3f5cb3cc9118 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 23cc444d-ff15-441b-8a0c-2ded49fd3cf0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202303] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 356a69b0-7898-440d-9473-7b4572ed2315 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 853.202303] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance b6e4f743-5692-4b46-892c-6c8917ccef98 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 8e0e13a7-eac5-4176-8536-2906d13c390e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Migration 05095ce1-1d9e-4bdc-8f8d-2b88b84020e3 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 853.202303] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 4f429be7-c594-4a72-aebd-0b746219231c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202726] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 41f6300d-7462-4e11-b32b-8c892e87bafc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202726] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance aebe41ad-496d-4a53-b023-c8df2bca04f8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202726] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 86e24eb9-6a55-4658-9e62-64713b0289d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202830] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance ecd08348-b22f-49c5-b13f-b187506e38d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.202865] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance d13673f2-e99f-400b-bcdc-b7c38e197610 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 853.303756] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229025, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.344539] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52555244-0778-de15-7023-dbd292da3b83, 'name': SearchDatastore_Task, 'duration_secs': 0.060178} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.344837] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.345097] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 853.345502] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a02348ee-9b78-46a4-a72c-30463aabfd30 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.353943] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 853.353943] env[61768]: value = "task-1229030" [ 853.353943] env[61768]: _type = "Task" [ 853.353943] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.362294] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229030, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.461770] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229028, 'name': Rename_Task, 'duration_secs': 0.156706} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.463074] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.463074] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b55a7133-c8c2-4f86-82c0-f39860ff159a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.469058] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 853.469058] env[61768]: value = "task-1229031" [ 853.469058] env[61768]: _type = "Task" [ 853.469058] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.483646] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229031, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.597162] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.597370] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.597562] env[61768]: DEBUG nova.network.neutron [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 853.597749] env[61768]: DEBUG nova.objects.instance [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'info_cache' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 853.692311] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 853.692571] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 853.692775] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleting the datastore file [datastore1] 41f6300d-7462-4e11-b32b-8c892e87bafc {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.693060] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-94527804-4e68-4384-9736-c24bd8aa2924 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.701354] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 853.701354] env[61768]: value = "task-1229032" [ 853.701354] env[61768]: _type = "Task" [ 853.701354] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.705889] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 853.706179] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance with task_state "unshelving" is not being actively managed by this compute host but has allocations referencing this compute node (f9f7149c-7c4e-44e7-8000-07a202159d2f): {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocations during the task state transition. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1780}} [ 853.714777] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229032, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.802584] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229025, 'name': Rename_Task, 'duration_secs': 1.178616} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.802900] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 853.803135] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29b7323f-f02d-4f86-a558-c00cd8f99dae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.812149] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 853.812149] env[61768]: value = "task-1229033" [ 853.812149] env[61768]: _type = "Task" [ 853.812149] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.822082] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229033, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.864939] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229030, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.993761] env[61768]: DEBUG oslo_vmware.api [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229031, 'name': PowerOnVM_Task, 'duration_secs': 0.508256} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.994152] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 853.994278] env[61768]: INFO nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Took 15.35 seconds to spawn the instance on the hypervisor. [ 853.994473] env[61768]: DEBUG nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 853.995345] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c9172e-3de0-44b2-9a82-1e31489149fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.103583] env[61768]: DEBUG nova.objects.base [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Object Instance<7697ecda-ef30-447d-a92a-3f5cb3cc9118> lazy-loaded attributes: flavor,info_cache {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 854.186103] env[61768]: DEBUG nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.210245] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 12f92761-45b6-49c6-96bd-1c0fa05521c6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 854.210642] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 854.210866] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=200GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 854.221374] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.221731] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.221968] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.222264] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.222521] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.222799] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.223170] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.223460] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.223771] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.224100] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.224337] env[61768]: DEBUG nova.virt.hardware [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.225072] env[61768]: DEBUG oslo_vmware.api [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229032, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.292653} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.226245] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be818846-2fbf-4f69-a135-0d714cb51592 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.230190] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 854.230580] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 854.230745] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 854.231202] env[61768]: INFO nova.compute.manager [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Took 2.76 seconds to destroy the instance on the hypervisor. [ 854.231394] env[61768]: DEBUG oslo.service.loopingcall [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.233178] env[61768]: DEBUG nova.compute.manager [-] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 854.233251] env[61768]: DEBUG nova.network.neutron [-] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 854.244868] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c59621a-99ee-486b-bce6-1d0e6908f452 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.327095] env[61768]: DEBUG oslo_vmware.api [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229033, 'name': PowerOnVM_Task, 'duration_secs': 0.486776} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.327443] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 854.327657] env[61768]: INFO nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Took 9.52 seconds to spawn the instance on the hypervisor. [ 854.327844] env[61768]: DEBUG nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.328628] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be63cef7-b54d-4661-92b0-087c98a60200 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.369148] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229030, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.514785} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.369148] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 854.369148] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 854.369148] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-55c7d50c-5add-41e8-9d19-fb61f3213e67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.376163] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 854.376163] env[61768]: value = "task-1229034" [ 854.376163] env[61768]: _type = "Task" [ 854.376163] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.392824] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229034, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.496144] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38a2dcf-711f-458a-a63a-8a8c59061a05 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.504669] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbc3e16b-3c91-41a0-93f2-e9ce27dd08ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.539667] env[61768]: INFO nova.compute.manager [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Took 44.28 seconds to build instance. [ 854.541232] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5585f0-f9c3-448b-a588-8d65dab4f292 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.550763] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e3f1af-d457-4dd9-a138-6019f7137b9a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.567766] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.665734] env[61768]: DEBUG nova.compute.manager [req-bfbdc330-f899-42bf-bb69-4c9c14e76b5b req-2742d0de-c657-4cf3-abb8-1e7615862dc9 service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Received event network-vif-plugged-be0213d6-1e3b-41cf-a6c5-b37ac80034f3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.665977] env[61768]: DEBUG oslo_concurrency.lockutils [req-bfbdc330-f899-42bf-bb69-4c9c14e76b5b req-2742d0de-c657-4cf3-abb8-1e7615862dc9 service nova] Acquiring lock "d13673f2-e99f-400b-bcdc-b7c38e197610-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.666243] env[61768]: DEBUG oslo_concurrency.lockutils [req-bfbdc330-f899-42bf-bb69-4c9c14e76b5b req-2742d0de-c657-4cf3-abb8-1e7615862dc9 service nova] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.666445] env[61768]: DEBUG oslo_concurrency.lockutils [req-bfbdc330-f899-42bf-bb69-4c9c14e76b5b req-2742d0de-c657-4cf3-abb8-1e7615862dc9 service nova] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.666699] env[61768]: DEBUG nova.compute.manager [req-bfbdc330-f899-42bf-bb69-4c9c14e76b5b req-2742d0de-c657-4cf3-abb8-1e7615862dc9 service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] No waiting events found dispatching network-vif-plugged-be0213d6-1e3b-41cf-a6c5-b37ac80034f3 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.666898] env[61768]: WARNING nova.compute.manager [req-bfbdc330-f899-42bf-bb69-4c9c14e76b5b req-2742d0de-c657-4cf3-abb8-1e7615862dc9 service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Received unexpected event network-vif-plugged-be0213d6-1e3b-41cf-a6c5-b37ac80034f3 for instance with vm_state building and task_state spawning. [ 854.743669] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "86e24eb9-6a55-4658-9e62-64713b0289d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.826623] env[61768]: DEBUG nova.network.neutron [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Successfully updated port: be0213d6-1e3b-41cf-a6c5-b37ac80034f3 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.851116] env[61768]: INFO nova.compute.manager [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Took 27.85 seconds to build instance. [ 854.886837] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229034, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.30671} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.888366] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 854.889227] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c694a42b-882b-4aaf-a8d7-467d269b4bb2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.912700] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 854.912946] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cafb6fa-b210-4376-bcc5-54d5db8c0113 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.928589] env[61768]: DEBUG nova.network.neutron [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating instance_info_cache with network_info: [{"id": "51b2319f-ceb1-41db-8019-1d918e70080d", "address": "fa:16:3e:7d:47:ab", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b2319f-ce", "ovs_interfaceid": "51b2319f-ceb1-41db-8019-1d918e70080d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.937521] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 854.937521] env[61768]: value = "task-1229035" [ 854.937521] env[61768]: _type = "Task" [ 854.937521] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.948294] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229035, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.045739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13ebcc7e-1f3d-4f2f-aae6-69ccf0b6d3ba tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.788s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.045739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.301s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.045739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "86e24eb9-6a55-4658-9e62-64713b0289d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.045739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.045739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.047924] env[61768]: INFO nova.compute.manager [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Terminating instance [ 855.049851] env[61768]: DEBUG nova.compute.manager [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 855.050066] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 855.050956] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de0f64c-6ef6-4d3b-b506-21592257d82e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.059699] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 855.059980] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9c305b5-abc1-430e-9b50-52040fbd62e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.066545] env[61768]: DEBUG oslo_vmware.api [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 855.066545] env[61768]: value = "task-1229036" [ 855.066545] env[61768]: _type = "Task" [ 855.066545] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.070140] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.079338] env[61768]: DEBUG oslo_vmware.api [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229036, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.219939] env[61768]: DEBUG nova.network.neutron [-] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.331363] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "refresh_cache-d13673f2-e99f-400b-bcdc-b7c38e197610" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.331363] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "refresh_cache-d13673f2-e99f-400b-bcdc-b7c38e197610" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.331363] env[61768]: DEBUG nova.network.neutron [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 855.352844] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc2657f-b038-422b-bae5-18ea6b9cff50 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.363s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.432626] env[61768]: DEBUG oslo_concurrency.lockutils [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.451960] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229035, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.577847] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 855.578137] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.424s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.578406] env[61768]: DEBUG oslo_vmware.api [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229036, 'name': PowerOffVM_Task, 'duration_secs': 0.306087} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.579079] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.353s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.582109] env[61768]: INFO nova.compute.claims [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.587485] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 855.587485] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 855.587485] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d7bd2c0b-88e6-495c-9bdb-4667453ad85e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.651746] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 855.651997] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 855.652212] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleting the datastore file [datastore1] 86e24eb9-6a55-4658-9e62-64713b0289d7 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 855.653991] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33b4e0a8-524a-4af3-a617-031253e73a81 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.666029] env[61768]: DEBUG oslo_vmware.api [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 855.666029] env[61768]: value = "task-1229038" [ 855.666029] env[61768]: _type = "Task" [ 855.666029] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.675834] env[61768]: DEBUG oslo_vmware.api [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229038, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.723185] env[61768]: INFO nova.compute.manager [-] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Took 1.49 seconds to deallocate network for instance. [ 855.906451] env[61768]: DEBUG nova.network.neutron [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.939281] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 855.939685] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8a306e7-05e2-4371-9585-9dae5c6de898 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.961158] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229035, 'name': ReconfigVM_Task, 'duration_secs': 0.527987} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.962895] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 855.964032] env[61768]: DEBUG oslo_vmware.api [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 855.964032] env[61768]: value = "task-1229039" [ 855.964032] env[61768]: _type = "Task" [ 855.964032] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.964032] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4153f112-103d-404b-8d0a-f2514fbcf7a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.976787] env[61768]: DEBUG oslo_vmware.api [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229039, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.992729] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 855.992729] env[61768]: value = "task-1229040" [ 855.992729] env[61768]: _type = "Task" [ 855.992729] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.006698] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229040, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.174847] env[61768]: DEBUG oslo_vmware.api [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229038, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130893} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.175279] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 856.175382] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 856.175582] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 856.175770] env[61768]: INFO nova.compute.manager [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 856.176054] env[61768]: DEBUG oslo.service.loopingcall [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.176271] env[61768]: DEBUG nova.compute.manager [-] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.176368] env[61768]: DEBUG nova.network.neutron [-] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 856.249149] env[61768]: DEBUG nova.network.neutron [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Updating instance_info_cache with network_info: [{"id": "be0213d6-1e3b-41cf-a6c5-b37ac80034f3", "address": "fa:16:3e:bd:3a:05", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe0213d6-1e", "ovs_interfaceid": "be0213d6-1e3b-41cf-a6c5-b37ac80034f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.269886] env[61768]: DEBUG nova.compute.manager [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.270108] env[61768]: DEBUG nova.compute.manager [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing instance network info cache due to event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.270334] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.270479] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.270641] env[61768]: DEBUG nova.network.neutron [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.305901] env[61768]: INFO nova.compute.manager [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Took 0.58 seconds to detach 1 volumes for instance. [ 856.479564] env[61768]: DEBUG oslo_vmware.api [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229039, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.507094] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229040, 'name': Rename_Task, 'duration_secs': 0.188641} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.507727] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 856.509013] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66e3c885-27d0-495b-a2a4-d010f8f582bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.516815] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 856.516815] env[61768]: value = "task-1229041" [ 856.516815] env[61768]: _type = "Task" [ 856.516815] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.537028] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229041, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.750271] env[61768]: DEBUG nova.compute.manager [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Received event network-vif-deleted-bc255400-bfd6-496f-a447-290139231203 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.750498] env[61768]: DEBUG nova.compute.manager [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Received event network-changed-be0213d6-1e3b-41cf-a6c5-b37ac80034f3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.750798] env[61768]: DEBUG nova.compute.manager [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Refreshing instance network info cache due to event network-changed-be0213d6-1e3b-41cf-a6c5-b37ac80034f3. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.750929] env[61768]: DEBUG oslo_concurrency.lockutils [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] Acquiring lock "refresh_cache-d13673f2-e99f-400b-bcdc-b7c38e197610" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.753536] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "refresh_cache-d13673f2-e99f-400b-bcdc-b7c38e197610" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.755898] env[61768]: DEBUG nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Instance network_info: |[{"id": "be0213d6-1e3b-41cf-a6c5-b37ac80034f3", "address": "fa:16:3e:bd:3a:05", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe0213d6-1e", "ovs_interfaceid": "be0213d6-1e3b-41cf-a6c5-b37ac80034f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 856.755898] env[61768]: DEBUG oslo_concurrency.lockutils [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] Acquired lock "refresh_cache-d13673f2-e99f-400b-bcdc-b7c38e197610" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.755898] env[61768]: DEBUG nova.network.neutron [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Refreshing network info cache for port be0213d6-1e3b-41cf-a6c5-b37ac80034f3 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 856.755898] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bd:3a:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c3291573-fad8-48cc-a965-c3554e7cee4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'be0213d6-1e3b-41cf-a6c5-b37ac80034f3', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.764110] env[61768]: DEBUG oslo.service.loopingcall [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.770864] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 856.773721] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4faefa0-7ba9-4906-89ee-dc80a422584e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.805248] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.805248] env[61768]: value = "task-1229042" [ 856.805248] env[61768]: _type = "Task" [ 856.805248] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.819821] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.829064] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229042, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.916820] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80bcffab-14a2-4322-a7d4-2dcbe23bfb57 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.926055] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb44c176-46a7-4594-bade-7da11d98a50f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.967656] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466508ab-ea56-4923-bf8c-02fcba83a537 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.981044] env[61768]: DEBUG oslo_vmware.api [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229039, 'name': PowerOnVM_Task, 'duration_secs': 0.603574} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.982724] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 856.982978] env[61768]: DEBUG nova.compute.manager [None req-56aa85cd-5e8f-4eb2-b261-b58ab9b5b040 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.984074] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffbd905-1f34-458f-be9e-a113e1b89f31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.987631] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab78c7a2-807f-4d0e-8de8-763c4d4c4f77 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.004178] env[61768]: DEBUG nova.compute.provider_tree [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.016153] env[61768]: DEBUG nova.network.neutron [-] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.033064] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229041, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.137405] env[61768]: DEBUG nova.network.neutron [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updated VIF entry in instance network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.137838] env[61768]: DEBUG nova.network.neutron [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.143018] env[61768]: DEBUG nova.network.neutron [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Updated VIF entry in instance network info cache for port be0213d6-1e3b-41cf-a6c5-b37ac80034f3. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 857.143362] env[61768]: DEBUG nova.network.neutron [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Updating instance_info_cache with network_info: [{"id": "be0213d6-1e3b-41cf-a6c5-b37ac80034f3", "address": "fa:16:3e:bd:3a:05", "network": {"id": "7af199c4-5f4e-40f7-9ff5-409dccf7c2ea", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-2112498317-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "06ce9d5af7e740ff882439e4486a0aad", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c3291573-fad8-48cc-a965-c3554e7cee4e", "external-id": "nsx-vlan-transportzone-115", "segmentation_id": 115, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbe0213d6-1e", "ovs_interfaceid": "be0213d6-1e3b-41cf-a6c5-b37ac80034f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.319310] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229042, 'name': CreateVM_Task, 'duration_secs': 0.352727} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.319559] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 857.320235] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.320407] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.320750] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 857.321032] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd76a54b-0c4d-4784-af2c-3cc45cb5467b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.326718] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 857.326718] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272d848-d0c7-d0d1-e147-93181a4fe3e4" [ 857.326718] env[61768]: _type = "Task" [ 857.326718] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.335610] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272d848-d0c7-d0d1-e147-93181a4fe3e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.351777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 857.352016] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.511737] env[61768]: DEBUG nova.scheduler.client.report [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.522757] env[61768]: INFO nova.compute.manager [-] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Took 1.35 seconds to deallocate network for instance. [ 857.527476] env[61768]: DEBUG oslo_vmware.api [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229041, 'name': PowerOnVM_Task, 'duration_secs': 0.805757} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.531341] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 857.531341] env[61768]: DEBUG nova.compute.manager [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.532143] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a17cc9-5caf-4e09-b5de-bd0f26352539 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.640935] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4e54b32-a960-46f2-bf21-159cdab20357 req-1018ff0c-b44b-4061-8fe4-344079032527 service nova] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.645636] env[61768]: DEBUG oslo_concurrency.lockutils [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] Releasing lock "refresh_cache-d13673f2-e99f-400b-bcdc-b7c38e197610" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.645877] env[61768]: DEBUG nova.compute.manager [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Received event network-vif-deleted-a8acff1f-7c19-4525-bc2d-3d6345758f75 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 857.646079] env[61768]: INFO nova.compute.manager [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Neutron deleted interface a8acff1f-7c19-4525-bc2d-3d6345758f75; detaching it from the instance and deleting it from the info cache [ 857.646264] env[61768]: DEBUG nova.network.neutron [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.837955] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272d848-d0c7-d0d1-e147-93181a4fe3e4, 'name': SearchDatastore_Task, 'duration_secs': 0.033317} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.838316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.838566] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.838889] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.839077] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.839271] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 857.839545] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51c884e3-8966-4131-aef9-8c585f51decb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.851620] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 857.851823] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 857.852598] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a386dde1-5376-4f18-91c1-730b092551ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.855009] env[61768]: DEBUG nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 857.861453] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 857.861453] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d60291-9673-2fd5-7f0b-342d421336ab" [ 857.861453] env[61768]: _type = "Task" [ 857.861453] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.870620] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d60291-9673-2fd5-7f0b-342d421336ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.018537] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.019096] env[61768]: DEBUG nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 858.022699] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.865s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.023471] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.026517] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.056s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.026867] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.029297] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.974s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.029506] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.031328] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 10.452s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.035038] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.050688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.064048] env[61768]: INFO nova.scheduler.client.report [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleted allocations for instance b6e4f743-5692-4b46-892c-6c8917ccef98 [ 858.066317] env[61768]: INFO nova.scheduler.client.report [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Deleted allocations for instance 356a69b0-7898-440d-9473-7b4572ed2315 [ 858.148674] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-50187a0f-7ae7-494a-b231-e893891df9f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.159469] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60d65b4e-52f3-4028-8f1a-aef69a097d3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.191839] env[61768]: DEBUG nova.compute.manager [req-a127b5e2-9235-462d-909f-d403b4357989 req-a1cf7c21-9eca-47c5-82c0-aba5fe707b0a service nova] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Detach interface failed, port_id=a8acff1f-7c19-4525-bc2d-3d6345758f75, reason: Instance 86e24eb9-6a55-4658-9e62-64713b0289d7 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 858.374455] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d60291-9673-2fd5-7f0b-342d421336ab, 'name': SearchDatastore_Task, 'duration_secs': 0.0326} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.375272] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5b31da7-09b4-4cb0-a6ec-ce6896006e4a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.380870] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 858.380870] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272dac1-838c-cd0a-76ea-6c53f8ddb0ba" [ 858.380870] env[61768]: _type = "Task" [ 858.380870] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.381933] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 858.392110] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272dac1-838c-cd0a-76ea-6c53f8ddb0ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.531678] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8bddf01f-20d5-46a6-aa10-2925ee483886 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 33.793s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.533698] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 11.155s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.534443] env[61768]: INFO nova.compute.manager [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Unshelving [ 858.538925] env[61768]: DEBUG nova.compute.utils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.540677] env[61768]: DEBUG nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.540851] env[61768]: DEBUG nova.network.neutron [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 858.577777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ca7e9acf-2d08-4346-ae74-a556ea92ebb7 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "356a69b0-7898-440d-9473-7b4572ed2315" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.522s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.581618] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f14a692e-fdca-43e7-a186-70462f116e76 tempest-ListServersNegativeTestJSON-1752101294 tempest-ListServersNegativeTestJSON-1752101294-project-member] Lock "b6e4f743-5692-4b46-892c-6c8917ccef98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.601s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.612298] env[61768]: DEBUG nova.policy [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c70625cf86045bc97589ed6b71e153c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e179026b20164ee2ba8ebd5e4784a0d1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.767700] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba6fc7e-080c-48f5-8d7b-91c4c07bd65c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.776298] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae063d5-8687-434b-813c-f6191dc1ec9d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.809706] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a3308c9-67dd-4d16-8bc1-caca0bd7386a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.818337] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db26f03-534d-4bad-a518-ec8ad3184daa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.833183] env[61768]: DEBUG nova.compute.provider_tree [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.895155] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5272dac1-838c-cd0a-76ea-6c53f8ddb0ba, 'name': SearchDatastore_Task, 'duration_secs': 0.010158} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.895452] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.895725] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] d13673f2-e99f-400b-bcdc-b7c38e197610/d13673f2-e99f-400b-bcdc-b7c38e197610.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 858.896014] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-216c8cbe-6f8b-4f06-8509-e15821898722 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.906218] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 858.906218] env[61768]: value = "task-1229043" [ 858.906218] env[61768]: _type = "Task" [ 858.906218] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.916192] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229043, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.921126] env[61768]: DEBUG nova.network.neutron [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Successfully created port: aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.981652] env[61768]: INFO nova.compute.manager [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Rebuilding instance [ 859.035033] env[61768]: DEBUG nova.compute.manager [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.036919] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f7f1b53-7433-4536-ba52-a633bfcb5979 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.045106] env[61768]: DEBUG nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 859.339489] env[61768]: DEBUG nova.scheduler.client.report [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.418567] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229043, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473546} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.418912] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] d13673f2-e99f-400b-bcdc-b7c38e197610/d13673f2-e99f-400b-bcdc-b7c38e197610.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 859.419168] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.419424] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3072b25-c33a-467d-9463-5fc553cdc5b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.427017] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 859.427017] env[61768]: value = "task-1229044" [ 859.427017] env[61768]: _type = "Task" [ 859.427017] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.435980] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229044, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.550896] env[61768]: INFO nova.virt.block_device [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Booting with volume 9ddc3f0e-f5f4-449a-878f-bda8467085c3 at /dev/sda [ 859.555328] env[61768]: DEBUG nova.compute.utils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 859.558082] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 859.558082] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08a65244-cf2e-4dd1-a6ae-08af6983bbc8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.566294] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 859.566294] env[61768]: value = "task-1229045" [ 859.566294] env[61768]: _type = "Task" [ 859.566294] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.580267] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229045, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.603975] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf3f9a2a-50dd-425f-b960-a52fd92ec1b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.622662] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c20f00e-d3f8-40b8-98c1-27f4d46ea794 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.655703] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d96a658d-0d25-4516-922e-55029dc3db45 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.666131] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85978d2b-554d-4312-b70c-d419b6cbd90f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.701045] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-049c419d-612a-4f35-97b4-95f36a5e2698 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.708964] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30537df5-3646-472f-93dd-782c8c69d1cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.724181] env[61768]: DEBUG nova.virt.block_device [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Updating existing volume attachment record: 57362d71-eaeb-4f72-a1ee-8bc42c4c7f32 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 859.939918] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229044, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.263098} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.940233] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.941129] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20be3e03-0972-4fc4-83ed-506d1cb19445 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.965021] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] d13673f2-e99f-400b-bcdc-b7c38e197610/d13673f2-e99f-400b-bcdc-b7c38e197610.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.965344] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-83e15e54-1418-4dae-bb34-2a5c7d036cd0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.986122] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 859.986122] env[61768]: value = "task-1229046" [ 859.986122] env[61768]: _type = "Task" [ 859.986122] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.996939] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229046, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.057763] env[61768]: INFO nova.virt.block_device [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Booting with volume b2dc03b4-7708-4dfc-ba53-f371792dd1ae at /dev/sdb [ 860.078046] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229045, 'name': PowerOffVM_Task, 'duration_secs': 0.249698} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.078380] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 860.078625] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 860.079512] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c41f78a-8a75-487d-a078-ff222061280c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.087727] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 860.088037] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e522a7fa-c5aa-43d0-8663-cbfc20472080 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.092622] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f97d335b-9653-4481-b881-019d297d6a10 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.102676] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a96a5c-ef73-4d92-9ba1-96d22ed7333d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.139172] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-928e6da4-ff37-43b0-aab5-69faed975945 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.150730] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f0ac6c-813e-44f7-9901-9c6743c16c11 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.187489] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227ff6ff-713e-4717-a8bb-c66910e2df10 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.190424] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 860.190634] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 860.190849] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleting the datastore file [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 860.191151] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57238bf6-28ec-48d9-bdb7-ba23e66c23a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.201145] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0910cbd0-471b-48e6-a586-018f1f24c5b4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.204304] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 860.204304] env[61768]: value = "task-1229048" [ 860.204304] env[61768]: _type = "Task" [ 860.204304] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.214138] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229048, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.217734] env[61768]: DEBUG nova.virt.block_device [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating existing volume attachment record: c8ecb62d-95ff-4d4c-a115-bf03b0d110f5 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 860.353621] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.322s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.356871] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.085s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.358533] env[61768]: INFO nova.compute.claims [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.503198] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229046, 'name': ReconfigVM_Task, 'duration_secs': 0.307786} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.503578] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Reconfigured VM instance instance-00000052 to attach disk [datastore1] d13673f2-e99f-400b-bcdc-b7c38e197610/d13673f2-e99f-400b-bcdc-b7c38e197610.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.505804] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0693778-f2e4-4b3f-99cc-7bbedab01078 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.513880] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 860.513880] env[61768]: value = "task-1229050" [ 860.513880] env[61768]: _type = "Task" [ 860.513880] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.522877] env[61768]: DEBUG nova.compute.manager [req-3601d4fb-c9bb-452b-a8cf-fa64a874e9a1 req-823e68fe-b536-4071-9e05-b508e01a8b9b service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Received event network-vif-plugged-aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.523252] env[61768]: DEBUG oslo_concurrency.lockutils [req-3601d4fb-c9bb-452b-a8cf-fa64a874e9a1 req-823e68fe-b536-4071-9e05-b508e01a8b9b service nova] Acquiring lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.523408] env[61768]: DEBUG oslo_concurrency.lockutils [req-3601d4fb-c9bb-452b-a8cf-fa64a874e9a1 req-823e68fe-b536-4071-9e05-b508e01a8b9b service nova] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.523628] env[61768]: DEBUG oslo_concurrency.lockutils [req-3601d4fb-c9bb-452b-a8cf-fa64a874e9a1 req-823e68fe-b536-4071-9e05-b508e01a8b9b service nova] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.524089] env[61768]: DEBUG nova.compute.manager [req-3601d4fb-c9bb-452b-a8cf-fa64a874e9a1 req-823e68fe-b536-4071-9e05-b508e01a8b9b service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] No waiting events found dispatching network-vif-plugged-aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.524256] env[61768]: WARNING nova.compute.manager [req-3601d4fb-c9bb-452b-a8cf-fa64a874e9a1 req-823e68fe-b536-4071-9e05-b508e01a8b9b service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Received unexpected event network-vif-plugged-aab2c242-227c-4cf0-9a64-75e0237360d7 for instance with vm_state building and task_state block_device_mapping. [ 860.531031] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229050, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.624024] env[61768]: DEBUG nova.network.neutron [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Successfully updated port: aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.715148] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229048, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.925292] env[61768]: INFO nova.scheduler.client.report [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted allocation for migration 05095ce1-1d9e-4bdc-8f8d-2b88b84020e3 [ 861.025683] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229050, 'name': Rename_Task, 'duration_secs': 0.163073} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.026244] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 861.026450] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-67895972-01a2-4e83-b36d-a9617aadae47 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.034914] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 861.034914] env[61768]: value = "task-1229053" [ 861.034914] env[61768]: _type = "Task" [ 861.034914] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.045797] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229053, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.126232] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquiring lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 861.126232] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquired lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 861.126429] env[61768]: DEBUG nova.network.neutron [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 861.217456] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229048, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.604501} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.218401] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 861.218401] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 861.218401] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 861.432353] env[61768]: DEBUG oslo_concurrency.lockutils [None req-609d5792-e586-4f70-9965-cd216aa2a689 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 17.331s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.547814] env[61768]: DEBUG oslo_vmware.api [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229053, 'name': PowerOnVM_Task, 'duration_secs': 0.488131} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.548160] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 861.548332] env[61768]: INFO nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Took 7.36 seconds to spawn the instance on the hypervisor. [ 861.548522] env[61768]: DEBUG nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.549417] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b998692e-cc82-4356-8a44-e98dcea9e3fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.619178] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1b631ff-b95a-4268-ab86-42d988ce9367 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.629332] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b78b30-e8c2-4846-b9c3-caafb1acf6a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.668289] env[61768]: DEBUG nova.network.neutron [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.670930] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da50d977-ea37-49fc-8b92-d93d306c1ba1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.679294] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48851f41-bdd3-4f64-8744-b8921b5bb3cd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.693443] env[61768]: DEBUG nova.compute.provider_tree [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.806601] env[61768]: DEBUG nova.network.neutron [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Updating instance_info_cache with network_info: [{"id": "aab2c242-227c-4cf0-9a64-75e0237360d7", "address": "fa:16:3e:5b:d4:e1", "network": {"id": "8cace282-2f8f-44f7-8efd-d2ad73d9bf49", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1445107616-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e179026b20164ee2ba8ebd5e4784a0d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab2c242-22", "ovs_interfaceid": "aab2c242-227c-4cf0-9a64-75e0237360d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.842544] env[61768]: DEBUG nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.843119] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.843353] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.843517] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.843704] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.843856] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.844023] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.844242] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.844408] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.844583] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.844756] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.844939] env[61768]: DEBUG nova.virt.hardware [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.845867] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76c37b38-30cb-4019-a01c-f877250aeeaa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.855061] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d53d2a-bb9b-444e-85e5-0f86a7d851d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.069589] env[61768]: INFO nova.compute.manager [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Took 19.96 seconds to build instance. [ 862.198416] env[61768]: DEBUG nova.scheduler.client.report [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 862.250880] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.251166] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.251333] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.251555] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.251744] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.251905] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.252135] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.252319] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.252494] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.252665] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.252920] env[61768]: DEBUG nova.virt.hardware [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.254183] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926b6e63-5103-4086-9d64-05a264eb0c66 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.266604] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d9a68f-8b87-4200-a55a-1e51ba3dd1e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.282287] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:c4:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ce058b2d-df85-481c-a996-cc179d534f1a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e2db4501-9ba5-4717-8ce1-2565bb20c603', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.289742] env[61768]: DEBUG oslo.service.loopingcall [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.290576] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 862.291112] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2bdb76e0-f256-448f-88ab-3d25042981b0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.308615] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Releasing lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.308950] env[61768]: DEBUG nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance network_info: |[{"id": "aab2c242-227c-4cf0-9a64-75e0237360d7", "address": "fa:16:3e:5b:d4:e1", "network": {"id": "8cace282-2f8f-44f7-8efd-d2ad73d9bf49", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1445107616-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e179026b20164ee2ba8ebd5e4784a0d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab2c242-22", "ovs_interfaceid": "aab2c242-227c-4cf0-9a64-75e0237360d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.309366] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:d4:e1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '0721b358-3768-472d-95f8-6d6755ab1635', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aab2c242-227c-4cf0-9a64-75e0237360d7', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.316652] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Creating folder: Project (e179026b20164ee2ba8ebd5e4784a0d1). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.318477] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3c706499-34a6-4760-9ab9-4ddc12681173 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.320107] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.320107] env[61768]: value = "task-1229054" [ 862.320107] env[61768]: _type = "Task" [ 862.320107] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.329673] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229054, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.333250] env[61768]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 862.333416] env[61768]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61768) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 862.333783] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Folder already exists: Project (e179026b20164ee2ba8ebd5e4784a0d1). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.333985] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Creating folder: Instances. Parent ref: group-v265460. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 862.334243] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-95562a18-82a9-4c8b-8e97-97a899f6f36f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.349634] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Created folder: Instances in parent group-v265460. [ 862.349997] env[61768]: DEBUG oslo.service.loopingcall [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.350501] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 862.350562] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91f82a3a-8e93-481b-94d8-693346b5a2e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.372333] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.372333] env[61768]: value = "task-1229057" [ 862.372333] env[61768]: _type = "Task" [ 862.372333] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.381456] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229057, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.492359] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "4f429be7-c594-4a72-aebd-0b746219231c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.492359] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.492359] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "4f429be7-c594-4a72-aebd-0b746219231c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.492359] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.492359] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.494379] env[61768]: INFO nova.compute.manager [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Terminating instance [ 862.498034] env[61768]: DEBUG nova.compute.manager [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.498433] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.500023] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5e1292-af1a-4a73-a592-a912423cfa76 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.512742] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.512742] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5627d59e-4b74-45ae-8edc-5b98a913c0e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.525255] env[61768]: DEBUG oslo_vmware.api [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 862.525255] env[61768]: value = "task-1229058" [ 862.525255] env[61768]: _type = "Task" [ 862.525255] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.534887] env[61768]: DEBUG oslo_vmware.api [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229058, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.571800] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13e4dffb-ed9c-4338-8a44-8aa487dfdda1 tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.473s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.705249] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.705956] env[61768]: DEBUG nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.710012] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.891s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.710335] env[61768]: DEBUG nova.objects.instance [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'resources' on Instance uuid 41f6300d-7462-4e11-b32b-8c892e87bafc {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.722636] env[61768]: DEBUG nova.compute.manager [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Received event network-changed-aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.722860] env[61768]: DEBUG nova.compute.manager [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Refreshing instance network info cache due to event network-changed-aab2c242-227c-4cf0-9a64-75e0237360d7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.723092] env[61768]: DEBUG oslo_concurrency.lockutils [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] Acquiring lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.723247] env[61768]: DEBUG oslo_concurrency.lockutils [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] Acquired lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.723416] env[61768]: DEBUG nova.network.neutron [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Refreshing network info cache for port aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 862.766546] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "d13673f2-e99f-400b-bcdc-b7c38e197610" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.767256] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.767256] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "d13673f2-e99f-400b-bcdc-b7c38e197610-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.767458] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.767543] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.769763] env[61768]: INFO nova.compute.manager [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Terminating instance [ 862.771740] env[61768]: DEBUG nova.compute.manager [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 862.771954] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 862.772900] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aab051a6-acbb-453b-8249-36bdbd904a98 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.782064] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.782332] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ad31a49-2d89-4b8b-a1c9-f64bc566bd14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.789053] env[61768]: DEBUG oslo_vmware.api [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 862.789053] env[61768]: value = "task-1229060" [ 862.789053] env[61768]: _type = "Task" [ 862.789053] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.797402] env[61768]: DEBUG oslo_vmware.api [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229060, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.831612] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229054, 'name': CreateVM_Task, 'duration_secs': 0.454215} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.831789] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.832580] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.832790] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.833164] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.833462] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-707a7a5d-bd3d-4208-8ea6-ce2072c87c6f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.838910] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 862.838910] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52176d69-82ad-cff8-e201-3afc238e42d9" [ 862.838910] env[61768]: _type = "Task" [ 862.838910] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.848353] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52176d69-82ad-cff8-e201-3afc238e42d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.883204] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229057, 'name': CreateVM_Task, 'duration_secs': 0.392743} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.883383] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 862.884102] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'device_type': None, 'guest_format': None, 'mount_device': '/dev/sda', 'attachment_id': '57362d71-eaeb-4f72-a1ee-8bc42c4c7f32', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265469', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'name': 'volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e', 'attached_at': '', 'detached_at': '', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'serial': '9ddc3f0e-f5f4-449a-878f-bda8467085c3'}, 'disk_bus': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=61768) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 862.884340] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Root volume attach. Driver type: vmdk {{(pid=61768) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 862.885454] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1caff6dd-a62a-4da0-b16a-ea2322cce827 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.894578] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc891b0-6342-490f-a08b-ada4c5efc402 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.902054] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efda69b4-2894-4cbe-ae07-0c0e88bde412 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.908512] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-c879bb94-bf18-44d2-a94b-1ed862eee676 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.919060] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 862.919060] env[61768]: value = "task-1229061" [ 862.919060] env[61768]: _type = "Task" [ 862.919060] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.929560] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.033649] env[61768]: DEBUG oslo_vmware.api [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229058, 'name': PowerOffVM_Task, 'duration_secs': 0.261042} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.033977] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.034175] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.034489] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b9872ad-13f5-4c2f-8b60-361400aa07b4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.063366] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.063714] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.104372] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.104603] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.104799] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleting the datastore file [datastore2] 4f429be7-c594-4a72-aebd-0b746219231c {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.105399] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72908570-38d0-4dca-ae79-b3f032c90273 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.112669] env[61768]: DEBUG oslo_vmware.api [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 863.112669] env[61768]: value = "task-1229063" [ 863.112669] env[61768]: _type = "Task" [ 863.112669] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.120893] env[61768]: DEBUG oslo_vmware.api [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229063, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.213925] env[61768]: DEBUG nova.compute.utils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 863.218782] env[61768]: DEBUG nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 863.219043] env[61768]: DEBUG nova.network.neutron [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 863.274455] env[61768]: DEBUG nova.policy [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fd5852bc16429aa2904b4351c8249d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7857b5217cc499ca8c9c3b8a5471db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 863.301375] env[61768]: DEBUG oslo_vmware.api [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229060, 'name': PowerOffVM_Task, 'duration_secs': 0.228496} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.302478] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 863.302478] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 863.302478] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c991ba01-7cda-47dd-bf53-9e5bece067ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.359028] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52176d69-82ad-cff8-e201-3afc238e42d9, 'name': SearchDatastore_Task, 'duration_secs': 0.011708} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.359028] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.359028] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.359028] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.359381] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.359662] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.360108] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7dbff731-6d8a-40c6-bf29-6a01ec47d116 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.376585] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 863.376585] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 863.376585] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleting the datastore file [datastore1] d13673f2-e99f-400b-bcdc-b7c38e197610 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.376585] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e71fb7f8-42a2-4f35-8d0b-a807eb611025 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.386994] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.389071] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 863.396072] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14e1e0a5-58f7-428f-b7ef-b6e605b84514 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.400397] env[61768]: DEBUG oslo_vmware.api [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for the task: (returnval){ [ 863.400397] env[61768]: value = "task-1229065" [ 863.400397] env[61768]: _type = "Task" [ 863.400397] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.406512] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 863.406512] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52efd896-acbf-d000-58d3-2ecdf5915997" [ 863.406512] env[61768]: _type = "Task" [ 863.406512] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.416567] env[61768]: DEBUG oslo_vmware.api [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229065, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.426538] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52efd896-acbf-d000-58d3-2ecdf5915997, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.436414] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 29%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.515857] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22cd6f56-2c30-4566-b81f-eecaad05e4fc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.519782] env[61768]: DEBUG nova.network.neutron [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Updated VIF entry in instance network info cache for port aab2c242-227c-4cf0-9a64-75e0237360d7. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 863.520273] env[61768]: DEBUG nova.network.neutron [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Updating instance_info_cache with network_info: [{"id": "aab2c242-227c-4cf0-9a64-75e0237360d7", "address": "fa:16:3e:5b:d4:e1", "network": {"id": "8cace282-2f8f-44f7-8efd-d2ad73d9bf49", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1445107616-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e179026b20164ee2ba8ebd5e4784a0d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab2c242-22", "ovs_interfaceid": "aab2c242-227c-4cf0-9a64-75e0237360d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.527732] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08d4e886-d5c2-427f-a1c4-88095007baf7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.568185] env[61768]: DEBUG nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 863.572066] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58113697-8067-4099-affd-4909d2161988 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.582882] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c013f3bc-2eb1-4730-ac61-857bb6e3bfb9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.588665] env[61768]: DEBUG nova.network.neutron [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Successfully created port: 3b83cae3-264d-4937-a2a1-7defcb2897bb {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.606711] env[61768]: DEBUG nova.compute.provider_tree [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.626422] env[61768]: DEBUG oslo_vmware.api [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229063, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.353293} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.626742] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.626968] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 863.627184] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 863.627372] env[61768]: INFO nova.compute.manager [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Took 1.13 seconds to destroy the instance on the hypervisor. [ 863.627628] env[61768]: DEBUG oslo.service.loopingcall [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.627828] env[61768]: DEBUG nova.compute.manager [-] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.627924] env[61768]: DEBUG nova.network.neutron [-] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 863.720696] env[61768]: DEBUG nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.914782] env[61768]: DEBUG oslo_vmware.api [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Task: {'id': task-1229065, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218755} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.915080] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.915284] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 863.915471] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 863.915658] env[61768]: INFO nova.compute.manager [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Took 1.14 seconds to destroy the instance on the hypervisor. [ 863.915973] env[61768]: DEBUG oslo.service.loopingcall [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.916656] env[61768]: DEBUG nova.compute.manager [-] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 863.916758] env[61768]: DEBUG nova.network.neutron [-] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 863.923422] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52efd896-acbf-d000-58d3-2ecdf5915997, 'name': SearchDatastore_Task, 'duration_secs': 0.023515} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.927954] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7896770-7fbf-4b05-84af-507d7122df92 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.936958] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 45%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.938672] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 863.938672] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524209fd-a4ba-2212-44ad-ba5a1b80cfd8" [ 863.938672] env[61768]: _type = "Task" [ 863.938672] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.948338] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524209fd-a4ba-2212-44ad-ba5a1b80cfd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.023823] env[61768]: DEBUG oslo_concurrency.lockutils [req-8233ef4e-13b4-4d36-a21e-4d520581ef34 req-69fa4e4a-d75d-4864-8e72-4e635989b9bc service nova] Releasing lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.025760] env[61768]: DEBUG nova.compute.manager [req-84b93241-c9fe-481e-bc53-a65f2b9bde23 req-ce010346-f5dd-47f7-bcd3-e3364ba1c007 service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Received event network-vif-deleted-0e23d603-ab56-4c07-b562-335a39497bac {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.025978] env[61768]: INFO nova.compute.manager [req-84b93241-c9fe-481e-bc53-a65f2b9bde23 req-ce010346-f5dd-47f7-bcd3-e3364ba1c007 service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Neutron deleted interface 0e23d603-ab56-4c07-b562-335a39497bac; detaching it from the instance and deleting it from the info cache [ 864.026291] env[61768]: DEBUG nova.network.neutron [req-84b93241-c9fe-481e-bc53-a65f2b9bde23 req-ce010346-f5dd-47f7-bcd3-e3364ba1c007 service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.101214] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.113389] env[61768]: DEBUG nova.scheduler.client.report [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 864.435586] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 58%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.454212] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524209fd-a4ba-2212-44ad-ba5a1b80cfd8, 'name': SearchDatastore_Task, 'duration_secs': 0.033182} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.454637] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.454824] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 864.455351] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7398ad61-8d96-4d12-9d89-d1acb50b2762 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.464742] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 864.464742] env[61768]: value = "task-1229066" [ 864.464742] env[61768]: _type = "Task" [ 864.464742] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.474340] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229066, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.497419] env[61768]: DEBUG nova.network.neutron [-] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.529956] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60054a90-ccc0-41c4-9057-9c524d3f405f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.543057] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4427762-50b8-4018-93ea-dc5b1b834e1e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.578571] env[61768]: DEBUG nova.compute.manager [req-84b93241-c9fe-481e-bc53-a65f2b9bde23 req-ce010346-f5dd-47f7-bcd3-e3364ba1c007 service nova] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Detach interface failed, port_id=0e23d603-ab56-4c07-b562-335a39497bac, reason: Instance 4f429be7-c594-4a72-aebd-0b746219231c could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 864.617496] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.907s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.620079] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.585s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.620350] env[61768]: DEBUG nova.objects.instance [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lazy-loading 'resources' on Instance uuid 86e24eb9-6a55-4658-9e62-64713b0289d7 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.639243] env[61768]: INFO nova.scheduler.client.report [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted allocations for instance 41f6300d-7462-4e11-b32b-8c892e87bafc [ 864.701037] env[61768]: DEBUG nova.network.neutron [-] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.729954] env[61768]: DEBUG nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.752015] env[61768]: DEBUG nova.compute.manager [req-019ec5b4-5664-4a66-8d94-542c072a57dd req-7502b9bb-baae-4802-aeb5-2b1b8985c625 service nova] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Received event network-vif-deleted-be0213d6-1e3b-41cf-a6c5-b37ac80034f3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.759432] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.760912] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.760912] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.760912] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.760912] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.760912] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.760912] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.760912] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.761317] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.761317] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.761384] env[61768]: DEBUG nova.virt.hardware [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.762719] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5b0d1a-4ff9-40d6-b2c4-3d6d52aa0af1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.775367] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ac4773-6315-4939-ba0c-ccb516b2ebd4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.937831] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 73%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.981420] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229066, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.000313] env[61768]: INFO nova.compute.manager [-] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Took 1.37 seconds to deallocate network for instance. [ 865.150507] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8884a9c8-f7b1-474c-93be-19f701c8a9cd tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "41f6300d-7462-4e11-b32b-8c892e87bafc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.688s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.203321] env[61768]: INFO nova.compute.manager [-] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Took 1.29 seconds to deallocate network for instance. [ 865.373273] env[61768]: DEBUG nova.network.neutron [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Successfully updated port: 3b83cae3-264d-4937-a2a1-7defcb2897bb {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.383701] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18fd4e4-9a8c-4a74-a947-5b6aa228a108 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.393770] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eacbf79-cc6b-4c45-898a-2582ddd1afc7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.438277] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dec4aed-91cf-4069-b74a-f6600fb45b67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.448424] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 88%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.452828] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e686e78-bbc5-4eef-98c5-d6c9e2f3f882 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.469404] env[61768]: DEBUG nova.compute.provider_tree [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 865.480667] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229066, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.632233} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.481681] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 865.481976] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 865.482262] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7d0c75cf-d0d3-4f48-aef9-cf52aa6d0c83 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.491976] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 865.491976] env[61768]: value = "task-1229067" [ 865.491976] env[61768]: _type = "Task" [ 865.491976] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.501646] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229067, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.509433] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.712175] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.842250] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 865.876185] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "refresh_cache-12f92761-45b6-49c6-96bd-1c0fa05521c6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.876407] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "refresh_cache-12f92761-45b6-49c6-96bd-1c0fa05521c6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.876548] env[61768]: DEBUG nova.network.neutron [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 865.948534] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 97%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.975349] env[61768]: DEBUG nova.scheduler.client.report [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 866.003833] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229067, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076911} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.004339] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.005407] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c974b98-2c6d-4818-b692-49b47b5f8396 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.031247] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.031442] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd5fd603-5e10-4e08-a129-e418a544695a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.053695] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 866.053695] env[61768]: value = "task-1229068" [ 866.053695] env[61768]: _type = "Task" [ 866.053695] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.066125] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229068, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.409676] env[61768]: DEBUG nova.network.neutron [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 866.444807] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task} progress is 98%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.480639] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.861s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.483065] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 8.432s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.483274] env[61768]: DEBUG nova.objects.instance [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 866.499024] env[61768]: INFO nova.scheduler.client.report [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted allocations for instance 86e24eb9-6a55-4658-9e62-64713b0289d7 [ 866.570154] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229068, 'name': ReconfigVM_Task, 'duration_secs': 0.457573} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.570508] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54/9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 866.571162] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1c4e3a32-f278-457c-a7e8-fa25438d04cd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.581693] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 866.581693] env[61768]: value = "task-1229069" [ 866.581693] env[61768]: _type = "Task" [ 866.581693] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.588280] env[61768]: DEBUG nova.network.neutron [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Updating instance_info_cache with network_info: [{"id": "3b83cae3-264d-4937-a2a1-7defcb2897bb", "address": "fa:16:3e:5c:e8:b2", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b83cae3-26", "ovs_interfaceid": "3b83cae3-264d-4937-a2a1-7defcb2897bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.594441] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229069, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.782035] env[61768]: DEBUG nova.compute.manager [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Received event network-vif-plugged-3b83cae3-264d-4937-a2a1-7defcb2897bb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.782333] env[61768]: DEBUG oslo_concurrency.lockutils [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] Acquiring lock "12f92761-45b6-49c6-96bd-1c0fa05521c6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.782535] env[61768]: DEBUG oslo_concurrency.lockutils [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.782729] env[61768]: DEBUG oslo_concurrency.lockutils [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.782911] env[61768]: DEBUG nova.compute.manager [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] No waiting events found dispatching network-vif-plugged-3b83cae3-264d-4937-a2a1-7defcb2897bb {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.783106] env[61768]: WARNING nova.compute.manager [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Received unexpected event network-vif-plugged-3b83cae3-264d-4937-a2a1-7defcb2897bb for instance with vm_state building and task_state spawning. [ 866.783282] env[61768]: DEBUG nova.compute.manager [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Received event network-changed-3b83cae3-264d-4937-a2a1-7defcb2897bb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.783444] env[61768]: DEBUG nova.compute.manager [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Refreshing instance network info cache due to event network-changed-3b83cae3-264d-4937-a2a1-7defcb2897bb. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.783617] env[61768]: DEBUG oslo_concurrency.lockutils [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] Acquiring lock "refresh_cache-12f92761-45b6-49c6-96bd-1c0fa05521c6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.944936] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229061, 'name': RelocateVM_Task, 'duration_secs': 3.808522} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.945721] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 866.945821] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265469', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'name': 'volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e', 'attached_at': '', 'detached_at': '', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'serial': '9ddc3f0e-f5f4-449a-878f-bda8467085c3'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 866.946676] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae51a72a-ee38-494a-ade6-621a6a955c18 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.963315] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87e2948-9e03-4684-ab11-f8b296712682 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.984323] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3/volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.984546] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7287c01e-aacd-47f0-b8ec-329bdbf1856e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.008282] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 867.008282] env[61768]: value = "task-1229070" [ 867.008282] env[61768]: _type = "Task" [ 867.008282] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.009245] env[61768]: DEBUG oslo_concurrency.lockutils [None req-437d6271-dc55-4713-a870-276ce15f85bd tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "86e24eb9-6a55-4658-9e62-64713b0289d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.964s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.019063] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229070, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.090604] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229069, 'name': Rename_Task, 'duration_secs': 0.162767} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.090910] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 867.091204] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c00605ce-d3dc-4981-b433-0469b1b0b9e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.095375] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "refresh_cache-12f92761-45b6-49c6-96bd-1c0fa05521c6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.095666] env[61768]: DEBUG nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Instance network_info: |[{"id": "3b83cae3-264d-4937-a2a1-7defcb2897bb", "address": "fa:16:3e:5c:e8:b2", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b83cae3-26", "ovs_interfaceid": "3b83cae3-264d-4937-a2a1-7defcb2897bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.095940] env[61768]: DEBUG oslo_concurrency.lockutils [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] Acquired lock "refresh_cache-12f92761-45b6-49c6-96bd-1c0fa05521c6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.096205] env[61768]: DEBUG nova.network.neutron [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Refreshing network info cache for port 3b83cae3-264d-4937-a2a1-7defcb2897bb {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 867.097427] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:e8:b2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3b83cae3-264d-4937-a2a1-7defcb2897bb', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.105286] env[61768]: DEBUG oslo.service.loopingcall [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.106901] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 867.107243] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 867.107243] env[61768]: value = "task-1229071" [ 867.107243] env[61768]: _type = "Task" [ 867.107243] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.108012] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b7a0584-e5b9-4cb1-9678-307a5cd7311b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.131116] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.131116] env[61768]: value = "task-1229072" [ 867.131116] env[61768]: _type = "Task" [ 867.131116] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.134904] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229071, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.143973] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229072, 'name': CreateVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.202042] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.202199] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.503761] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ff24a402-5dc4-4b8d-8879-fab098d1325c tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.021s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.504921] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.123s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.506419] env[61768]: INFO nova.compute.claims [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 867.519364] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229070, 'name': ReconfigVM_Task, 'duration_secs': 0.245757} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.519539] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Reconfigured VM instance instance-00000053 to attach disk [datastore2] volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3/volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.524478] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-838558a0-37fc-4dd7-8a1d-1b09a7d49cc7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.539435] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 867.539435] env[61768]: value = "task-1229073" [ 867.539435] env[61768]: _type = "Task" [ 867.539435] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.547487] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229073, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.635440] env[61768]: DEBUG oslo_vmware.api [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229071, 'name': PowerOnVM_Task, 'duration_secs': 0.492823} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.638772] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 867.639088] env[61768]: DEBUG nova.compute.manager [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.641950] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f886a0b-d9b2-4b70-b9a0-59b80e0630c9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.649804] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229072, 'name': CreateVM_Task, 'duration_secs': 0.326128} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.651342] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 867.654557] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.654638] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.654947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.655448] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-128cba32-2298-4354-b76c-e2c6236e5094 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.659769] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 867.659769] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52affae5-9762-a766-f299-203794827c66" [ 867.659769] env[61768]: _type = "Task" [ 867.659769] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.668058] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52affae5-9762-a766-f299-203794827c66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.704698] env[61768]: DEBUG nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 867.879344] env[61768]: DEBUG nova.network.neutron [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Updated VIF entry in instance network info cache for port 3b83cae3-264d-4937-a2a1-7defcb2897bb. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 867.879715] env[61768]: DEBUG nova.network.neutron [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Updating instance_info_cache with network_info: [{"id": "3b83cae3-264d-4937-a2a1-7defcb2897bb", "address": "fa:16:3e:5c:e8:b2", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3b83cae3-26", "ovs_interfaceid": "3b83cae3-264d-4937-a2a1-7defcb2897bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.051641] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229073, 'name': ReconfigVM_Task, 'duration_secs': 0.125637} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.051641] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265469', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'name': 'volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e', 'attached_at': '', 'detached_at': '', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'serial': '9ddc3f0e-f5f4-449a-878f-bda8467085c3'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 868.051641] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-32c26c33-9da9-4a21-a7b7-1a09e676f79a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.060674] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 868.060674] env[61768]: value = "task-1229074" [ 868.060674] env[61768]: _type = "Task" [ 868.060674] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.067340] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229074, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.168137] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.174276] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52affae5-9762-a766-f299-203794827c66, 'name': SearchDatastore_Task, 'duration_secs': 0.009837} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.174677] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.174975] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.175301] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.175904] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.175904] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.176081] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0065d9c1-b25f-4f8d-8533-80151c611a94 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.186169] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.186370] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 868.187601] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1712d25e-6959-4f2e-87c3-f0c3d4ed3c22 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.193543] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 868.193543] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5202a772-9b1d-ed2e-8e4b-2ec5883cb2e4" [ 868.193543] env[61768]: _type = "Task" [ 868.193543] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.202520] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5202a772-9b1d-ed2e-8e4b-2ec5883cb2e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.215484] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.215788] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.216190] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.216478] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.216702] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.221925] env[61768]: INFO nova.compute.manager [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Terminating instance [ 868.223868] env[61768]: DEBUG nova.compute.manager [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.224093] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.225054] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa82abdd-b313-453b-bb13-388c76d76f84 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.233934] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.235254] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.235561] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08c08828-758d-4751-a28e-d216f7f26e6b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.242148] env[61768]: DEBUG oslo_vmware.api [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 868.242148] env[61768]: value = "task-1229075" [ 868.242148] env[61768]: _type = "Task" [ 868.242148] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.252226] env[61768]: DEBUG oslo_vmware.api [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229075, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.386080] env[61768]: DEBUG oslo_concurrency.lockutils [req-5665fd69-bc81-422f-b0cd-3eec540bec4f req-5e701aa7-0ea8-41aa-8dce-9e499557aa67 service nova] Releasing lock "refresh_cache-12f92761-45b6-49c6-96bd-1c0fa05521c6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.567688] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229074, 'name': Rename_Task, 'duration_secs': 0.179724} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.568092] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 868.568401] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39ce3373-b0fe-41df-ac4d-95dccedcfc99 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.574913] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 868.574913] env[61768]: value = "task-1229076" [ 868.574913] env[61768]: _type = "Task" [ 868.574913] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.584563] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229076, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.707488] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5202a772-9b1d-ed2e-8e4b-2ec5883cb2e4, 'name': SearchDatastore_Task, 'duration_secs': 0.020387} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.707488] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a171ad29-8eb9-4c41-975a-798426755ef6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.713753] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 868.713753] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528cc47c-9b1b-a355-6915-119cabb53fff" [ 868.713753] env[61768]: _type = "Task" [ 868.713753] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.722078] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528cc47c-9b1b-a355-6915-119cabb53fff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.737662] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-314c98cc-d3b3-4890-8675-7b74d17ca0af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.752594] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.752923] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.755429] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd818dbc-6234-4de5-91c4-102db2ac20b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.764378] env[61768]: DEBUG oslo_vmware.api [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229075, 'name': PowerOffVM_Task, 'duration_secs': 0.237044} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.790576] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 868.790845] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 868.791837] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5772acc3-48d7-4d5e-9ac1-2bcfe6624bb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.794103] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c7ba2f-3900-402e-86bc-de15fb319519 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.804442] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6f7176-48be-4b29-9498-0769e8952cd2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.818554] env[61768]: DEBUG nova.compute.provider_tree [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 868.855894] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 868.856323] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 868.856629] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleting the datastore file [datastore2] 3d96e47d-fc09-439e-bf0e-e60fc87ae958 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 868.857048] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed6f00d7-ba69-4e14-9ad2-4adedca9c5f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.864120] env[61768]: DEBUG oslo_vmware.api [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 868.864120] env[61768]: value = "task-1229078" [ 868.864120] env[61768]: _type = "Task" [ 868.864120] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.873168] env[61768]: DEBUG oslo_vmware.api [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229078, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.901435] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "aebe41ad-496d-4a53-b023-c8df2bca04f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.901777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.901926] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "aebe41ad-496d-4a53-b023-c8df2bca04f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.902163] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.902383] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.904930] env[61768]: INFO nova.compute.manager [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Terminating instance [ 868.907306] env[61768]: DEBUG nova.compute.manager [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 868.907534] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 868.908388] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd914cf6-b146-41be-9794-c38597b8c97f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.915953] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 868.916217] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-367b765f-ee3a-4caa-bd24-f752f102bdec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.921604] env[61768]: DEBUG oslo_vmware.api [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 868.921604] env[61768]: value = "task-1229079" [ 868.921604] env[61768]: _type = "Task" [ 868.921604] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.929132] env[61768]: DEBUG oslo_vmware.api [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229079, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.085900] env[61768]: DEBUG oslo_vmware.api [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229076, 'name': PowerOnVM_Task, 'duration_secs': 0.452043} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.086237] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 869.086465] env[61768]: INFO nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Took 7.24 seconds to spawn the instance on the hypervisor. [ 869.086656] env[61768]: DEBUG nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.087867] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-513bbe57-3e02-4a03-8867-0b806e708b54 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.223626] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528cc47c-9b1b-a355-6915-119cabb53fff, 'name': SearchDatastore_Task, 'duration_secs': 0.009368} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.223896] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.224189] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 12f92761-45b6-49c6-96bd-1c0fa05521c6/12f92761-45b6-49c6-96bd-1c0fa05521c6.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 869.224457] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af2e77f8-0741-4bb0-8562-ec1752a16026 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.230428] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 869.230428] env[61768]: value = "task-1229080" [ 869.230428] env[61768]: _type = "Task" [ 869.230428] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.238128] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229080, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.259885] env[61768]: DEBUG nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 869.339334] env[61768]: ERROR nova.scheduler.client.report [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [req-e399383c-615e-47bd-9645-6a39215b903f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e399383c-615e-47bd-9645-6a39215b903f"}]} [ 869.362261] env[61768]: DEBUG nova.scheduler.client.report [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 869.373496] env[61768]: DEBUG oslo_vmware.api [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159576} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.373756] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 869.373948] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 869.374152] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 869.374326] env[61768]: INFO nova.compute.manager [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Took 1.15 seconds to destroy the instance on the hypervisor. [ 869.374571] env[61768]: DEBUG oslo.service.loopingcall [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 869.374764] env[61768]: DEBUG nova.compute.manager [-] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 869.374861] env[61768]: DEBUG nova.network.neutron [-] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 869.381969] env[61768]: DEBUG nova.scheduler.client.report [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 869.382225] env[61768]: DEBUG nova.compute.provider_tree [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.394781] env[61768]: DEBUG nova.scheduler.client.report [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 869.413880] env[61768]: DEBUG nova.scheduler.client.report [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 869.433213] env[61768]: DEBUG oslo_vmware.api [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229079, 'name': PowerOffVM_Task, 'duration_secs': 0.220697} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.433393] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 869.433567] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 869.436291] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-87f3722d-7c25-4054-9e15-7d3f4678d95b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.506466] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 869.506769] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 869.507032] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleting the datastore file [datastore1] aebe41ad-496d-4a53-b023-c8df2bca04f8 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 869.507295] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7c40d7bb-6cd9-4d4f-bf3c-768b28c1d981 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.513974] env[61768]: DEBUG oslo_vmware.api [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 869.513974] env[61768]: value = "task-1229082" [ 869.513974] env[61768]: _type = "Task" [ 869.513974] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.522526] env[61768]: DEBUG oslo_vmware.api [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229082, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.606213] env[61768]: INFO nova.compute.manager [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Took 26.40 seconds to build instance. [ 869.688601] env[61768]: DEBUG nova.compute.manager [req-09c28e52-42b3-4941-9de7-84d37e9d7257 req-4a692389-22ff-4fa4-9eee-b232f5ab6734 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Received event network-vif-deleted-61d36258-9772-43d5-aa74-55c86e1596a7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 869.688813] env[61768]: INFO nova.compute.manager [req-09c28e52-42b3-4941-9de7-84d37e9d7257 req-4a692389-22ff-4fa4-9eee-b232f5ab6734 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Neutron deleted interface 61d36258-9772-43d5-aa74-55c86e1596a7; detaching it from the instance and deleting it from the info cache [ 869.689028] env[61768]: DEBUG nova.network.neutron [req-09c28e52-42b3-4941-9de7-84d37e9d7257 req-4a692389-22ff-4fa4-9eee-b232f5ab6734 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.698376] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f6bf55f-b5b0-4f41-9e3e-bfde3d0c0c3a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.707071] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c74a69d-177d-4142-a875-c43f8e13a971 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.739465] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553a4b89-0933-4f34-b935-49fa007f7084 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.750092] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68701fa6-a7e1-414f-8341-96ea3c811a8a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.753751] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229080, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.766373] env[61768]: DEBUG nova.compute.provider_tree [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.780994] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.031058] env[61768]: DEBUG oslo_vmware.api [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229082, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.417704} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.031058] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 870.031058] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 870.031058] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 870.031058] env[61768]: INFO nova.compute.manager [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Took 1.12 seconds to destroy the instance on the hypervisor. [ 870.031058] env[61768]: DEBUG oslo.service.loopingcall [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 870.031058] env[61768]: DEBUG nova.compute.manager [-] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 870.031058] env[61768]: DEBUG nova.network.neutron [-] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 870.108909] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8d06006-6952-4618-a6db-ec954d6155b5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.914s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.165336] env[61768]: DEBUG nova.network.neutron [-] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.191304] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b30a493d-3c78-444e-a918-268c733b3c32 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.201179] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3418d90-f3e1-4e28-a947-f6e8b11cace1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.232649] env[61768]: DEBUG nova.compute.manager [req-09c28e52-42b3-4941-9de7-84d37e9d7257 req-4a692389-22ff-4fa4-9eee-b232f5ab6734 service nova] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Detach interface failed, port_id=61d36258-9772-43d5-aa74-55c86e1596a7, reason: Instance 3d96e47d-fc09-439e-bf0e-e60fc87ae958 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 870.244907] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229080, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.820436} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.246353] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 12f92761-45b6-49c6-96bd-1c0fa05521c6/12f92761-45b6-49c6-96bd-1c0fa05521c6.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 870.246866] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.247077] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8db9c78-a680-4d74-80b6-e8b1a83badf1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.253711] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 870.253711] env[61768]: value = "task-1229083" [ 870.253711] env[61768]: _type = "Task" [ 870.253711] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.261423] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229083, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.307368] env[61768]: DEBUG nova.scheduler.client.report [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Updated inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with generation 103 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 870.307951] env[61768]: DEBUG nova.compute.provider_tree [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 103 to 104 during operation: update_inventory {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 870.308203] env[61768]: DEBUG nova.compute.provider_tree [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 870.668683] env[61768]: INFO nova.compute.manager [-] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Took 1.29 seconds to deallocate network for instance. [ 870.763672] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229083, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073074} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.763911] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.764773] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e7b116-8396-4eda-a177-2783c09f527d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.786511] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] 12f92761-45b6-49c6-96bd-1c0fa05521c6/12f92761-45b6-49c6-96bd-1c0fa05521c6.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.786805] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-738ce5a7-237d-4722-ba12-12de0f976d8c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.806862] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 870.806862] env[61768]: value = "task-1229084" [ 870.806862] env[61768]: _type = "Task" [ 870.806862] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.817172] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.312s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.817686] env[61768]: DEBUG nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 870.820344] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229084, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.820826] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.720s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.822293] env[61768]: INFO nova.compute.claims [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.872203] env[61768]: DEBUG nova.network.neutron [-] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.176044] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.317365] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229084, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.322813] env[61768]: DEBUG nova.compute.utils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 871.324559] env[61768]: DEBUG nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 871.324559] env[61768]: DEBUG nova.network.neutron [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 871.368141] env[61768]: DEBUG nova.policy [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 871.374604] env[61768]: INFO nova.compute.manager [-] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Took 1.35 seconds to deallocate network for instance. [ 871.665904] env[61768]: DEBUG nova.network.neutron [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Successfully created port: 604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 871.719212] env[61768]: DEBUG nova.compute.manager [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Received event network-vif-deleted-7758e5b9-f72b-49f5-84df-7ddcc681d518 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.719353] env[61768]: DEBUG nova.compute.manager [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Received event network-changed-aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.719605] env[61768]: DEBUG nova.compute.manager [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Refreshing instance network info cache due to event network-changed-aab2c242-227c-4cf0-9a64-75e0237360d7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 871.719811] env[61768]: DEBUG oslo_concurrency.lockutils [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] Acquiring lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.719928] env[61768]: DEBUG oslo_concurrency.lockutils [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] Acquired lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.720092] env[61768]: DEBUG nova.network.neutron [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Refreshing network info cache for port aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 871.817630] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229084, 'name': ReconfigVM_Task, 'duration_secs': 0.985968} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.817911] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Reconfigured VM instance instance-00000054 to attach disk [datastore2] 12f92761-45b6-49c6-96bd-1c0fa05521c6/12f92761-45b6-49c6-96bd-1c0fa05521c6.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.818619] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-73dc9db0-7920-4c50-86a2-fa2b2b72c4ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.824694] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 871.824694] env[61768]: value = "task-1229085" [ 871.824694] env[61768]: _type = "Task" [ 871.824694] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.827932] env[61768]: DEBUG nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 871.844962] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229085, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.884017] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.108677] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88e33f42-58d4-44fb-a802-7207711be27b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.116137] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-120e6db0-5f3d-49eb-9ea5-0155dc0cb2f8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.147531] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9ff261-0853-4b2c-a18a-379caf07453e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.154982] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d14b9b55-3ba1-403f-9ae2-569c5e281dbe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.169759] env[61768]: DEBUG nova.compute.provider_tree [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 872.333715] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229085, 'name': Rename_Task, 'duration_secs': 0.276821} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.334045] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 872.334196] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c2a8cd6-5854-4275-abf6-b547481df4fb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.343206] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 872.343206] env[61768]: value = "task-1229086" [ 872.343206] env[61768]: _type = "Task" [ 872.343206] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.350596] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229086, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.434342] env[61768]: DEBUG nova.network.neutron [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Updated VIF entry in instance network info cache for port aab2c242-227c-4cf0-9a64-75e0237360d7. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 872.434727] env[61768]: DEBUG nova.network.neutron [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Updating instance_info_cache with network_info: [{"id": "aab2c242-227c-4cf0-9a64-75e0237360d7", "address": "fa:16:3e:5b:d4:e1", "network": {"id": "8cace282-2f8f-44f7-8efd-d2ad73d9bf49", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-1445107616-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.179", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e179026b20164ee2ba8ebd5e4784a0d1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "0721b358-3768-472d-95f8-6d6755ab1635", "external-id": "nsx-vlan-transportzone-314", "segmentation_id": 314, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaab2c242-22", "ovs_interfaceid": "aab2c242-227c-4cf0-9a64-75e0237360d7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.673064] env[61768]: DEBUG nova.scheduler.client.report [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.846091] env[61768]: DEBUG nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 872.857144] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229086, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.870455] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 872.870693] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 872.870857] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 872.871056] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 872.871217] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 872.871379] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 872.871594] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 872.871761] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 872.871935] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 872.872120] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 872.872381] env[61768]: DEBUG nova.virt.hardware [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 872.873248] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c67ae1b-f2b5-423a-abfd-5ce92272ce43 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.880270] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e429d5-eacc-4661-ba9c-ea7e0a571940 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.937632] env[61768]: DEBUG oslo_concurrency.lockutils [req-bf66ee95-bfbb-4c96-835d-406e42567aa5 req-38717cc3-fecc-4a3c-a27e-b12666a4405e service nova] Releasing lock "refresh_cache-c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.178484] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.179377] env[61768]: DEBUG nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 873.181940] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.673s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.182162] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.188361] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.472s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.188625] env[61768]: DEBUG nova.objects.instance [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lazy-loading 'resources' on Instance uuid d13673f2-e99f-400b-bcdc-b7c38e197610 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.211719] env[61768]: INFO nova.scheduler.client.report [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted allocations for instance 4f429be7-c594-4a72-aebd-0b746219231c [ 873.306221] env[61768]: DEBUG nova.network.neutron [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Successfully updated port: 604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 873.357359] env[61768]: DEBUG oslo_vmware.api [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229086, 'name': PowerOnVM_Task, 'duration_secs': 0.954082} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.357713] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 873.357840] env[61768]: INFO nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Took 8.63 seconds to spawn the instance on the hypervisor. [ 873.358034] env[61768]: DEBUG nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.358795] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2b093d-30b6-4ebe-b95c-a9c4d4c902a0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.691960] env[61768]: DEBUG nova.compute.utils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.696647] env[61768]: DEBUG nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.697079] env[61768]: DEBUG nova.network.neutron [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 873.719407] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c93fd50-a370-4967-a9e4-414ed87d7bf4 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "4f429be7-c594-4a72-aebd-0b746219231c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.228s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.741948] env[61768]: DEBUG nova.policy [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9d7f72b4c22452d89004ca70d165e67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f4804c086b449608ef34ea22c018907', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.745714] env[61768]: DEBUG nova.compute.manager [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-vif-plugged-604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.746223] env[61768]: DEBUG oslo_concurrency.lockutils [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] Acquiring lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.746223] env[61768]: DEBUG oslo_concurrency.lockutils [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.746351] env[61768]: DEBUG oslo_concurrency.lockutils [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.746526] env[61768]: DEBUG nova.compute.manager [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] No waiting events found dispatching network-vif-plugged-604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 873.746806] env[61768]: WARNING nova.compute.manager [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received unexpected event network-vif-plugged-604a3820-78ee-48e4-bb9e-af1273963ff8 for instance with vm_state building and task_state spawning. [ 873.746877] env[61768]: DEBUG nova.compute.manager [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.747068] env[61768]: DEBUG nova.compute.manager [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing instance network info cache due to event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.747274] env[61768]: DEBUG oslo_concurrency.lockutils [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.747422] env[61768]: DEBUG oslo_concurrency.lockutils [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.747584] env[61768]: DEBUG nova.network.neutron [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing network info cache for port 604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 873.810396] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.875900] env[61768]: INFO nova.compute.manager [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Took 22.62 seconds to build instance. [ 873.884999] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7182d55a-75f5-4ac1-98fb-402e85e8d018 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.893546] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a03f7f7-07f4-499a-b56b-2939acfe753e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.925017] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40900b46-1884-4bd1-925d-6bda86d6cc5c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.932789] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fa6f59-5a7d-4565-b6ce-6e3566c12c4a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.946921] env[61768]: DEBUG nova.compute.provider_tree [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.975971] env[61768]: DEBUG nova.network.neutron [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Successfully created port: 5b4c6981-f46a-4e25-9f9a-51106a3988ac {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.197568] env[61768]: DEBUG nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 874.379160] env[61768]: DEBUG oslo_concurrency.lockutils [None req-508ce969-ef60-45f5-a97e-a1a598278205 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.136s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.380174] env[61768]: DEBUG nova.network.neutron [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.449655] env[61768]: DEBUG nova.scheduler.client.report [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.554500] env[61768]: DEBUG nova.network.neutron [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.954376] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.770s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.957027] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.115s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.957027] env[61768]: DEBUG nova.objects.instance [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'pci_requests' on Instance uuid 62cddff5-d499-4ef3-869d-3cdb0328640d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 874.977385] env[61768]: INFO nova.scheduler.client.report [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Deleted allocations for instance d13673f2-e99f-400b-bcdc-b7c38e197610 [ 875.057772] env[61768]: DEBUG oslo_concurrency.lockutils [req-16725159-5244-47fb-b929-003801f8d465 req-98b03843-f514-4f71-aa3c-e09eded2fede service nova] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.058149] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.058310] env[61768]: DEBUG nova.network.neutron [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.184386] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "12f92761-45b6-49c6-96bd-1c0fa05521c6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.184671] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.184893] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "12f92761-45b6-49c6-96bd-1c0fa05521c6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.185102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.185288] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.188111] env[61768]: INFO nova.compute.manager [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Terminating instance [ 875.190206] env[61768]: DEBUG nova.compute.manager [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 875.190531] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 875.191451] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e69e92-8c9d-4610-a635-67150b62009b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.199485] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 875.200020] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0007cc8f-e115-49fb-a19d-9eb3323056e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.206150] env[61768]: DEBUG nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 875.209733] env[61768]: DEBUG oslo_vmware.api [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 875.209733] env[61768]: value = "task-1229087" [ 875.209733] env[61768]: _type = "Task" [ 875.209733] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.217665] env[61768]: DEBUG oslo_vmware.api [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.238364] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 875.238634] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 875.238799] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 875.239267] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 875.239267] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 875.239392] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 875.239625] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 875.239799] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 875.240011] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 875.240221] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 875.240436] env[61768]: DEBUG nova.virt.hardware [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 875.241624] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa2bd75-4411-40d9-8599-9521a1ef1c9e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.249507] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2963653c-27c4-47d0-8c35-615bf8e66521 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.463648] env[61768]: DEBUG nova.objects.instance [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'numa_topology' on Instance uuid 62cddff5-d499-4ef3-869d-3cdb0328640d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.484397] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d71f1b97-4d55-4430-a8cb-0f50172413ae tempest-ImagesOneServerNegativeTestJSON-1643911918 tempest-ImagesOneServerNegativeTestJSON-1643911918-project-member] Lock "d13673f2-e99f-400b-bcdc-b7c38e197610" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.717s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.566117] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "8e0e13a7-eac5-4176-8536-2906d13c390e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.566117] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.592748] env[61768]: DEBUG nova.network.neutron [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.724072] env[61768]: DEBUG oslo_vmware.api [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229087, 'name': PowerOffVM_Task, 'duration_secs': 0.197732} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.724662] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 875.724977] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 875.725326] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a0c8d0de-508c-4669-8b37-07e23f98dde0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.863717] env[61768]: DEBUG nova.network.neutron [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.888491] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 875.888789] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 875.888963] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleting the datastore file [datastore2] 12f92761-45b6-49c6-96bd-1c0fa05521c6 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 875.889259] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fd990bc3-0c1f-4c1f-ab17-78591671246a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.895538] env[61768]: DEBUG oslo_vmware.api [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 875.895538] env[61768]: value = "task-1229089" [ 875.895538] env[61768]: _type = "Task" [ 875.895538] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.903192] env[61768]: DEBUG oslo_vmware.api [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229089, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.958524] env[61768]: DEBUG nova.network.neutron [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Successfully updated port: 5b4c6981-f46a-4e25-9f9a-51106a3988ac {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.963526] env[61768]: DEBUG nova.compute.manager [req-594f560a-aec9-4b98-995d-3c565fcd284c req-bf8ce0b0-a5d9-4c1b-9238-2e9caebeed61 service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Received event network-vif-plugged-5b4c6981-f46a-4e25-9f9a-51106a3988ac {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.963526] env[61768]: DEBUG oslo_concurrency.lockutils [req-594f560a-aec9-4b98-995d-3c565fcd284c req-bf8ce0b0-a5d9-4c1b-9238-2e9caebeed61 service nova] Acquiring lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.963526] env[61768]: DEBUG oslo_concurrency.lockutils [req-594f560a-aec9-4b98-995d-3c565fcd284c req-bf8ce0b0-a5d9-4c1b-9238-2e9caebeed61 service nova] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.963526] env[61768]: DEBUG oslo_concurrency.lockutils [req-594f560a-aec9-4b98-995d-3c565fcd284c req-bf8ce0b0-a5d9-4c1b-9238-2e9caebeed61 service nova] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.963526] env[61768]: DEBUG nova.compute.manager [req-594f560a-aec9-4b98-995d-3c565fcd284c req-bf8ce0b0-a5d9-4c1b-9238-2e9caebeed61 service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] No waiting events found dispatching network-vif-plugged-5b4c6981-f46a-4e25-9f9a-51106a3988ac {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 875.963526] env[61768]: WARNING nova.compute.manager [req-594f560a-aec9-4b98-995d-3c565fcd284c req-bf8ce0b0-a5d9-4c1b-9238-2e9caebeed61 service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Received unexpected event network-vif-plugged-5b4c6981-f46a-4e25-9f9a-51106a3988ac for instance with vm_state building and task_state spawning. [ 875.965756] env[61768]: INFO nova.compute.claims [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 876.070102] env[61768]: DEBUG nova.compute.utils [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.367286] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.367969] env[61768]: DEBUG nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Instance network_info: |[{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 876.368128] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:65:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd8383707-f093-40a7-a5ba-31b0e07cac45', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '604a3820-78ee-48e4-bb9e-af1273963ff8', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 876.375820] env[61768]: DEBUG oslo.service.loopingcall [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.377099] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 876.377099] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b5d5a4f-5d8b-4913-ac25-aa46ba435d05 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.396726] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.396726] env[61768]: value = "task-1229090" [ 876.396726] env[61768]: _type = "Task" [ 876.396726] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.407012] env[61768]: DEBUG oslo_vmware.api [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229089, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.129757} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.410290] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.410290] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 876.410467] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 876.410651] env[61768]: INFO nova.compute.manager [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Took 1.22 seconds to destroy the instance on the hypervisor. [ 876.410898] env[61768]: DEBUG oslo.service.loopingcall [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 876.411076] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229090, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.411286] env[61768]: DEBUG nova.compute.manager [-] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 876.411408] env[61768]: DEBUG nova.network.neutron [-] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 876.464209] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.464555] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.464621] env[61768]: DEBUG nova.network.neutron [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 876.573978] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.729133] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d0dfd6-716c-45f8-b391-155d5e64c7fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.737516] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a577cab1-a2f5-4228-83a3-e3c807ade322 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.772200] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9fdd30-3469-4e8d-abc7-8ba0d6af4077 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.783023] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050c33ec-40e1-46d1-a867-62cbe7a21984 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.793956] env[61768]: DEBUG nova.compute.provider_tree [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.911786] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229090, 'name': CreateVM_Task, 'duration_secs': 0.340627} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.912162] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 876.913040] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.913423] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.915020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.915020] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1d8e12d6-b344-480c-a8b5-078a945dcbe5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.918854] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 876.918854] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52986404-037a-c138-4236-2408195ea2c9" [ 876.918854] env[61768]: _type = "Task" [ 876.918854] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.927535] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52986404-037a-c138-4236-2408195ea2c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.000447] env[61768]: DEBUG nova.network.neutron [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.142132] env[61768]: DEBUG nova.network.neutron [-] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.158256] env[61768]: DEBUG nova.network.neutron [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance_info_cache with network_info: [{"id": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "address": "fa:16:3e:2b:a2:c7", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4c6981-f4", "ovs_interfaceid": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.297722] env[61768]: DEBUG nova.scheduler.client.report [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.430214] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52986404-037a-c138-4236-2408195ea2c9, 'name': SearchDatastore_Task, 'duration_secs': 0.008804} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.431024] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.431024] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 877.431174] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.431755] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.431755] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 877.431755] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a10f2ff7-08b9-4dc5-acae-21ac95a598a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.442636] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 877.442815] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 877.444042] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-129efc61-6316-4799-8c84-c72ba65cb7d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.449748] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 877.449748] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e41c10-c8a8-babd-f31e-046a8ba0b40f" [ 877.449748] env[61768]: _type = "Task" [ 877.449748] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.458550] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e41c10-c8a8-babd-f31e-046a8ba0b40f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.641348] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "8e0e13a7-eac5-4176-8536-2906d13c390e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.641690] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.641926] env[61768]: INFO nova.compute.manager [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Attaching volume d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b to /dev/sdb [ 877.645359] env[61768]: INFO nova.compute.manager [-] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Took 1.23 seconds to deallocate network for instance. [ 877.660903] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.661207] env[61768]: DEBUG nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Instance network_info: |[{"id": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "address": "fa:16:3e:2b:a2:c7", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4c6981-f4", "ovs_interfaceid": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 877.661606] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2b:a2:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b4c6981-f46a-4e25-9f9a-51106a3988ac', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 877.669329] env[61768]: DEBUG oslo.service.loopingcall [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 877.669449] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 877.669849] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c09cae54-31e9-4c06-9728-79bf84475b3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.690189] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7e638d-560d-42ac-85f6-82db233ac852 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.699315] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b524aa89-4540-471d-9b68-0287d061f7c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.701795] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 877.701795] env[61768]: value = "task-1229091" [ 877.701795] env[61768]: _type = "Task" [ 877.701795] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.710594] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229091, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.716221] env[61768]: DEBUG nova.virt.block_device [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Updating existing volume attachment record: bdb46c25-f30b-49d5-b7d4-1df44a74f8e0 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 877.805347] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.849s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.807568] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.640s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.807802] env[61768]: DEBUG nova.objects.instance [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 877.848977] env[61768]: INFO nova.network.neutron [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 877.960181] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e41c10-c8a8-babd-f31e-046a8ba0b40f, 'name': SearchDatastore_Task, 'duration_secs': 0.010762} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.961079] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7314a6f8-1f91-4fca-94d3-e79c95c9e560 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.968260] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 877.968260] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52825851-f957-9c39-2e6d-0245fa51b74d" [ 877.968260] env[61768]: _type = "Task" [ 877.968260] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.975221] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52825851-f957-9c39-2e6d-0245fa51b74d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.009882] env[61768]: DEBUG nova.compute.manager [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Received event network-changed-5b4c6981-f46a-4e25-9f9a-51106a3988ac {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.010124] env[61768]: DEBUG nova.compute.manager [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Refreshing instance network info cache due to event network-changed-5b4c6981-f46a-4e25-9f9a-51106a3988ac. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.010426] env[61768]: DEBUG oslo_concurrency.lockutils [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] Acquiring lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.010632] env[61768]: DEBUG oslo_concurrency.lockutils [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] Acquired lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.010833] env[61768]: DEBUG nova.network.neutron [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Refreshing network info cache for port 5b4c6981-f46a-4e25-9f9a-51106a3988ac {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.153390] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.212149] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229091, 'name': CreateVM_Task, 'duration_secs': 0.332621} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.212362] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 878.213169] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.213415] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.213848] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.214259] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5923d0ba-6e19-415a-8504-ac857af3083f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.219288] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 878.219288] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52423696-d58e-a682-9c2a-6948f046f890" [ 878.219288] env[61768]: _type = "Task" [ 878.219288] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.230341] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52423696-d58e-a682-9c2a-6948f046f890, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.476611] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52825851-f957-9c39-2e6d-0245fa51b74d, 'name': SearchDatastore_Task, 'duration_secs': 0.018057} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.476975] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.477655] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 5cc705a5-1423-43e1-8356-9abd67a8da3a/5cc705a5-1423-43e1-8356-9abd67a8da3a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 878.477780] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6014a536-fb88-4ba8-b529-25285c472b2d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.484431] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 878.484431] env[61768]: value = "task-1229095" [ 878.484431] env[61768]: _type = "Task" [ 878.484431] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.492367] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229095, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.731876] env[61768]: DEBUG nova.network.neutron [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updated VIF entry in instance network info cache for port 5b4c6981-f46a-4e25-9f9a-51106a3988ac. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 878.732287] env[61768]: DEBUG nova.network.neutron [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance_info_cache with network_info: [{"id": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "address": "fa:16:3e:2b:a2:c7", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4c6981-f4", "ovs_interfaceid": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.733749] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52423696-d58e-a682-9c2a-6948f046f890, 'name': SearchDatastore_Task, 'duration_secs': 0.011072} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.733971] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.734254] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 878.734512] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.734710] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.734902] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 878.735537] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe59f26c-5097-4ed4-81b8-a3a19cf28b31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.750139] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 878.750455] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 878.751952] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4748559b-2cef-4a22-98e6-9ca5ba9255cb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.759055] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 878.759055] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b888aa-3a7d-1bb2-f655-5c0ab702c6a2" [ 878.759055] env[61768]: _type = "Task" [ 878.759055] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.769537] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b888aa-3a7d-1bb2-f655-5c0ab702c6a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.818579] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e3caae2e-fbba-4c03-b6ea-bb90b8a49a0d tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.820717] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.585s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.822274] env[61768]: INFO nova.compute.claims [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.994421] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229095, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.43899} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.994714] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 5cc705a5-1423-43e1-8356-9abd67a8da3a/5cc705a5-1423-43e1-8356-9abd67a8da3a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 878.995071] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 878.995448] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71cb6e34-2ead-4846-803b-85e82ac36f9f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.005443] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 879.005443] env[61768]: value = "task-1229096" [ 879.005443] env[61768]: _type = "Task" [ 879.005443] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.017363] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229096, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.236944] env[61768]: DEBUG oslo_concurrency.lockutils [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] Releasing lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.237300] env[61768]: DEBUG nova.compute.manager [req-f3ce17fe-ee4d-48b9-a24e-97160b5f1500 req-5e09bace-8544-4b83-95e2-7251211b6efb service nova] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Received event network-vif-deleted-3b83cae3-264d-4937-a2a1-7defcb2897bb {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.271665] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b888aa-3a7d-1bb2-f655-5c0ab702c6a2, 'name': SearchDatastore_Task, 'duration_secs': 0.05609} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.271897] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb4250c5-9ee1-4535-9f27-d4477f360539 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.277225] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 879.277225] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523bb093-a539-7b7a-7640-b04aebbe36cd" [ 879.277225] env[61768]: _type = "Task" [ 879.277225] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.284957] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523bb093-a539-7b7a-7640-b04aebbe36cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.515731] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229096, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074172} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.515999] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 879.516912] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c5651a-9c35-4093-9963-1e200953f7e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.551194] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] 5cc705a5-1423-43e1-8356-9abd67a8da3a/5cc705a5-1423-43e1-8356-9abd67a8da3a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 879.551194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.551194] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.551971] env[61768]: DEBUG nova.network.neutron [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.553158] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07a6dde3-38af-49c7-abeb-7cf31093ab11 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.577496] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 879.577496] env[61768]: value = "task-1229097" [ 879.577496] env[61768]: _type = "Task" [ 879.577496] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.588392] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229097, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.788231] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523bb093-a539-7b7a-7640-b04aebbe36cd, 'name': SearchDatastore_Task, 'duration_secs': 0.009189} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.788514] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.788780] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] e2b7cb22-475b-4b1b-b30e-6061e888691e/e2b7cb22-475b-4b1b-b30e-6061e888691e.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 879.789051] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8f21f65b-fb4f-43e5-a6ca-3bd1808eda92 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.791919] env[61768]: DEBUG nova.network.neutron [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.799345] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 879.799345] env[61768]: value = "task-1229098" [ 879.799345] env[61768]: _type = "Task" [ 879.799345] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.807438] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229098, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.039215] env[61768]: DEBUG nova.compute.manager [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-vif-plugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.039510] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.039767] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.040761] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.042166] env[61768]: DEBUG nova.compute.manager [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] No waiting events found dispatching network-vif-plugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.042570] env[61768]: WARNING nova.compute.manager [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received unexpected event network-vif-plugged-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 for instance with vm_state shelved_offloaded and task_state spawning. [ 880.042827] env[61768]: DEBUG nova.compute.manager [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.043371] env[61768]: DEBUG nova.compute.manager [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing instance network info cache due to event network-changed-44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 880.043780] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.070092] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba7cc90-099a-4f24-aab3-d2812379e0de {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.083665] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e9af7e-5cb7-4502-8cdb-bad4f6e700b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.093620] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229097, 'name': ReconfigVM_Task, 'duration_secs': 0.279085} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.122288] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Reconfigured VM instance instance-00000055 to attach disk [datastore2] 5cc705a5-1423-43e1-8356-9abd67a8da3a/5cc705a5-1423-43e1-8356-9abd67a8da3a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.123473] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5fb356e4-9cbc-4045-88d5-222f0f1de2d6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.126389] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0277e08-e5a8-498d-bf7d-8e14eeb1030b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.137885] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbee3f9-d511-496e-8934-0e6937540eb0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.142266] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 880.142266] env[61768]: value = "task-1229099" [ 880.142266] env[61768]: _type = "Task" [ 880.142266] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.154329] env[61768]: DEBUG nova.compute.provider_tree [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 880.162023] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229099, 'name': Rename_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.294377] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.296932] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.297188] env[61768]: DEBUG nova.network.neutron [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Refreshing network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 880.308735] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229098, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445136} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.308998] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] e2b7cb22-475b-4b1b-b30e-6061e888691e/e2b7cb22-475b-4b1b-b30e-6061e888691e.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 880.310280] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 880.311131] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cb5758d-3825-4170-b725-fe7c6e71a3d8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.321996] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 880.321996] env[61768]: value = "task-1229101" [ 880.321996] env[61768]: _type = "Task" [ 880.321996] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.332023] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229101, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.334883] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='1f000fe2d76daa3639f8f73362eaa81e',container_format='bare',created_at=2024-10-09T19:57:49Z,direct_url=,disk_format='vmdk',id=1e83b86d-089f-4b79-940c-ef76d051d4f3,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-197007024-shelved',owner='aa61ecc36bc14769a2fc909ee95716ae',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-10-09T19:58:03Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 880.335182] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 880.335393] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 880.335637] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 880.336155] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 880.336155] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 880.336258] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 880.336435] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 880.336574] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 880.336742] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 880.336954] env[61768]: DEBUG nova.virt.hardware [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 880.338527] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970d9ab8-c0d8-44b1-aaf3-21d1690a2f79 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.346892] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62649d91-7118-4859-a913-972034d8d475 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.363842] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:98:a8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37434b93-dfdc-4a3f-bf5a-9f2cbe25a754', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '44b2acbd-7a9f-416b-b3b4-e877de5bfa85', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.371725] env[61768]: DEBUG oslo.service.loopingcall [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.372084] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 880.372557] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2e2cdd50-e8e2-4c5c-865a-17bbbe8ecb77 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.392638] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.392638] env[61768]: value = "task-1229102" [ 880.392638] env[61768]: _type = "Task" [ 880.392638] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.401219] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229102, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.652995] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229099, 'name': Rename_Task, 'duration_secs': 0.225389} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.653578] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 880.653967] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ff932d0-1e8b-4a88-8354-ec9fb66073f7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.658773] env[61768]: DEBUG nova.scheduler.client.report [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.663613] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 880.663613] env[61768]: value = "task-1229103" [ 880.663613] env[61768]: _type = "Task" [ 880.663613] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.672615] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229103, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.834237] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229101, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071333} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.834237] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 880.834237] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b076d2-dc88-460d-83a9-3b59d8fd2ea0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.858396] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] e2b7cb22-475b-4b1b-b30e-6061e888691e/e2b7cb22-475b-4b1b-b30e-6061e888691e.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 880.862441] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-74874122-3d53-4cd5-87ca-06bb7b05b3c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.878875] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 880.878875] env[61768]: value = "task-1229104" [ 880.878875] env[61768]: _type = "Task" [ 880.878875] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.891184] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229104, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.903192] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229102, 'name': CreateVM_Task, 'duration_secs': 0.378622} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.905927] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 880.906661] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.906854] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.907318] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 880.907975] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f44e8d66-5124-4c80-ad33-e03f94efd655 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.915031] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 880.915031] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c738a3-3580-540f-ed02-a1d64f128ff3" [ 880.915031] env[61768]: _type = "Task" [ 880.915031] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.922045] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c738a3-3580-540f-ed02-a1d64f128ff3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.140323] env[61768]: DEBUG nova.network.neutron [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updated VIF entry in instance network info cache for port 44b2acbd-7a9f-416b-b3b4-e877de5bfa85. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 881.140707] env[61768]: DEBUG nova.network.neutron [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.165344] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.165907] env[61768]: DEBUG nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 881.170289] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.388s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.170289] env[61768]: INFO nova.compute.claims [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.182725] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229103, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.391941] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229104, 'name': ReconfigVM_Task, 'duration_secs': 0.261496} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.392337] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Reconfigured VM instance instance-00000056 to attach disk [datastore2] e2b7cb22-475b-4b1b-b30e-6061e888691e/e2b7cb22-475b-4b1b-b30e-6061e888691e.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 881.393071] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-02a5d8a0-ec4d-4820-9a19-5fc243f424b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.400689] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 881.400689] env[61768]: value = "task-1229105" [ 881.400689] env[61768]: _type = "Task" [ 881.400689] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.413237] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229105, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.425996] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.426332] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Processing image 1e83b86d-089f-4b79-940c-ef76d051d4f3 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 881.426620] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3/1e83b86d-089f-4b79-940c-ef76d051d4f3.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.426889] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3/1e83b86d-089f-4b79-940c-ef76d051d4f3.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.427154] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 881.427461] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f4af0cfa-60a6-4397-8280-277b3fd59535 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.437810] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 881.438054] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 881.438791] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e5493ca-11d3-4a54-aaf9-db250df6afa0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.443988] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 881.443988] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a37aa0-c4bf-cdb5-ed8d-5b1448a28055" [ 881.443988] env[61768]: _type = "Task" [ 881.443988] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.451840] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a37aa0-c4bf-cdb5-ed8d-5b1448a28055, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.647923] env[61768]: DEBUG oslo_concurrency.lockutils [req-b4dacac0-cd90-4e53-b0af-cc5b31908543 req-cc686085-f1fb-4a5c-9967-5708a238596d service nova] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.684441] env[61768]: DEBUG nova.compute.utils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.685981] env[61768]: DEBUG oslo_vmware.api [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229103, 'name': PowerOnVM_Task, 'duration_secs': 0.570786} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.688017] env[61768]: DEBUG nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 881.688017] env[61768]: DEBUG nova.network.neutron [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 881.691633] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 881.691633] env[61768]: INFO nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Took 8.84 seconds to spawn the instance on the hypervisor. [ 881.691633] env[61768]: DEBUG nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.691633] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac9c2e18-dfc5-4e19-937b-92466507c58c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.788191] env[61768]: DEBUG nova.policy [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd1aa7c174648888b58f6a6df6f82a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfce0cf4a5d14042a929dac65aeb2ae3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.912181] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229105, 'name': Rename_Task, 'duration_secs': 0.171336} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.912843] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 881.913205] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f74f145e-8bba-473b-9771-590cf30b790b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.919834] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 881.919834] env[61768]: value = "task-1229106" [ 881.919834] env[61768]: _type = "Task" [ 881.919834] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.928655] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229106, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.957027] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Preparing fetch location {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 881.957027] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Fetch image to [datastore2] OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0/OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0.vmdk {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 881.957027] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Downloading stream optimized image 1e83b86d-089f-4b79-940c-ef76d051d4f3 to [datastore2] OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0/OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0.vmdk on the data store datastore2 as vApp {{(pid=61768) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 881.957027] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Downloading image file data 1e83b86d-089f-4b79-940c-ef76d051d4f3 to the ESX as VM named 'OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0' {{(pid=61768) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 882.042040] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 882.042040] env[61768]: value = "resgroup-9" [ 882.042040] env[61768]: _type = "ResourcePool" [ 882.042040] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 882.042544] env[61768]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-77ff4957-471d-45de-83bd-2395c77b4726 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.069119] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease: (returnval){ [ 882.069119] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b716fd-d84a-42e9-9680-ea7bbc715dc3" [ 882.069119] env[61768]: _type = "HttpNfcLease" [ 882.069119] env[61768]: } obtained for vApp import into resource pool (val){ [ 882.069119] env[61768]: value = "resgroup-9" [ 882.069119] env[61768]: _type = "ResourcePool" [ 882.069119] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 882.069429] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the lease: (returnval){ [ 882.069429] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b716fd-d84a-42e9-9680-ea7bbc715dc3" [ 882.069429] env[61768]: _type = "HttpNfcLease" [ 882.069429] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 882.076597] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 882.076597] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b716fd-d84a-42e9-9680-ea7bbc715dc3" [ 882.076597] env[61768]: _type = "HttpNfcLease" [ 882.076597] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 882.116687] env[61768]: DEBUG nova.network.neutron [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Successfully created port: c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.191121] env[61768]: DEBUG nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 882.216627] env[61768]: INFO nova.compute.manager [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Took 23.86 seconds to build instance. [ 882.278644] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 882.278927] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265490', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'name': 'volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e0e13a7-eac5-4176-8536-2906d13c390e', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'serial': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 882.279944] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d49547-c45b-4637-aac8-aa9787b1d67a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.303679] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d031e3b9-c56d-4992-b074-5f54aab25bab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.337052] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b/volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.342972] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e48845f3-ae36-47ff-a028-6ea3e16f892f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.365150] env[61768]: DEBUG oslo_vmware.api [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 882.365150] env[61768]: value = "task-1229108" [ 882.365150] env[61768]: _type = "Task" [ 882.365150] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.378654] env[61768]: DEBUG oslo_vmware.api [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229108, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.432964] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229106, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.535123] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9cbd8f8-d4f7-47e2-be5f-69a8051f2a78 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.542941] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0b1f25-4e88-40ac-97f2-e18f7c605727 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.579345] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04f0919-6608-4cf6-bbcb-bc7b437703d1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.585349] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 882.585349] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b716fd-d84a-42e9-9680-ea7bbc715dc3" [ 882.585349] env[61768]: _type = "HttpNfcLease" [ 882.585349] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 882.587442] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 882.587442] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b716fd-d84a-42e9-9680-ea7bbc715dc3" [ 882.587442] env[61768]: _type = "HttpNfcLease" [ 882.587442] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 882.588215] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1144bc9-3579-4911-91c9-7c76322722cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.593118] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e265ff0d-0cdc-471a-86b6-18f0bdbaf6e8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.605732] env[61768]: DEBUG nova.compute.provider_tree [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.610355] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b18d2-7218-ae49-e37b-c9ab58879c5f/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 882.610355] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b18d2-7218-ae49-e37b-c9ab58879c5f/disk-0.vmdk. {{(pid=61768) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 882.677268] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5e3b72fa-e487-4cd2-b80e-4aaaa8665ca6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.721090] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b42f9e38-ebef-4e96-aa31-9f841136ad8a tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.368s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.877509] env[61768]: DEBUG oslo_vmware.api [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229108, 'name': ReconfigVM_Task, 'duration_secs': 0.486081} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.880622] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Reconfigured VM instance instance-0000004d to attach disk [datastore2] volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b/volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 882.890408] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6cb0921-13bc-48c1-ac3e-2719fc698bf1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.915033] env[61768]: DEBUG oslo_vmware.api [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 882.915033] env[61768]: value = "task-1229112" [ 882.915033] env[61768]: _type = "Task" [ 882.915033] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.934967] env[61768]: DEBUG oslo_vmware.api [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229112, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.941882] env[61768]: DEBUG oslo_vmware.api [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229106, 'name': PowerOnVM_Task, 'duration_secs': 0.572719} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.944657] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 882.944657] env[61768]: INFO nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Took 7.74 seconds to spawn the instance on the hypervisor. [ 882.944657] env[61768]: DEBUG nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.945569] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc0e97a-fa6f-4902-ad59-7352c4d985d1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.147017] env[61768]: DEBUG nova.scheduler.client.report [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updated inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with generation 104 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 883.147446] env[61768]: DEBUG nova.compute.provider_tree [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 104 to 105 during operation: update_inventory {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 883.147753] env[61768]: DEBUG nova.compute.provider_tree [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 883.201160] env[61768]: DEBUG nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 883.224936] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 883.225265] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 883.225471] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 883.225699] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 883.225959] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 883.226182] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 883.226450] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 883.226683] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 883.226900] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 883.227109] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 883.227289] env[61768]: DEBUG nova.virt.hardware [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 883.228208] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef1b1a4-a5c2-4834-91ed-435c6cc1337f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.241807] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a087d4-6d62-4606-a7df-3ccac828989e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.423525] env[61768]: DEBUG oslo_vmware.api [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229112, 'name': ReconfigVM_Task, 'duration_secs': 0.196006} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.423967] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265490', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'name': 'volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e0e13a7-eac5-4176-8536-2906d13c390e', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'serial': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 883.467107] env[61768]: INFO nova.compute.manager [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Took 19.39 seconds to build instance. [ 883.646094] env[61768]: DEBUG nova.compute.manager [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.646325] env[61768]: DEBUG nova.compute.manager [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing instance network info cache due to event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.646555] env[61768]: DEBUG oslo_concurrency.lockutils [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.646711] env[61768]: DEBUG oslo_concurrency.lockutils [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.646882] env[61768]: DEBUG nova.network.neutron [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 883.655458] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.655957] env[61768]: DEBUG nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.659340] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.483s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.663264] env[61768]: DEBUG nova.objects.instance [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lazy-loading 'resources' on Instance uuid 3d96e47d-fc09-439e-bf0e-e60fc87ae958 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.844112] env[61768]: DEBUG nova.network.neutron [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Successfully updated port: c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.969653] env[61768]: DEBUG oslo_concurrency.lockutils [None req-15bd8e5e-82fd-46e6-92a0-d4479801961f tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.905s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.072504] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Completed reading data from the image iterator. {{(pid=61768) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 884.073683] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b18d2-7218-ae49-e37b-c9ab58879c5f/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 884.074478] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df4d0b76-d8c8-418b-b5fc-53090b6e1f50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.084088] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b18d2-7218-ae49-e37b-c9ab58879c5f/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 884.084376] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b18d2-7218-ae49-e37b-c9ab58879c5f/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 884.084681] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-207339ba-44b6-46f7-8086-f022094a7fc4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.160674] env[61768]: DEBUG nova.compute.utils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 884.162177] env[61768]: DEBUG nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 884.162396] env[61768]: DEBUG nova.network.neutron [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 884.267508] env[61768]: DEBUG oslo_vmware.rw_handles [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/527b18d2-7218-ae49-e37b-c9ab58879c5f/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 884.267660] env[61768]: INFO nova.virt.vmwareapi.images [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Downloaded image file data 1e83b86d-089f-4b79-940c-ef76d051d4f3 [ 884.268447] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f87a35a-ef7e-430c-849c-1b9f810350ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.292412] env[61768]: DEBUG nova.policy [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f7f7389d526c4ce9bc8ecf062b9921e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '180da0bcd5954eb2bc89a20cf8f42bb5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.293819] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-91c49d8a-959a-4b43-a9cf-a7d63fb94249 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.321784] env[61768]: INFO nova.virt.vmwareapi.images [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] The imported VM was unregistered [ 884.324820] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Caching image {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 884.325081] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating directory with path [datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 884.327721] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-403a8c8c-2831-4709-b9cb-b265e4279817 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.350960] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.352640] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.353124] env[61768]: DEBUG nova.network.neutron [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 884.365311] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created directory with path [datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3 {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 884.367033] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0/OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0.vmdk to [datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3/1e83b86d-089f-4b79-940c-ef76d051d4f3.vmdk. {{(pid=61768) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 884.367033] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e132b493-c68d-4e30-b7ea-57762a8020c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.385065] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 884.385065] env[61768]: value = "task-1229114" [ 884.385065] env[61768]: _type = "Task" [ 884.385065] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.408907] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.466802] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e365e28f-e8c0-4c70-89be-544d9f7fd514 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.474501] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39010fc-73ff-4fce-85b5-f849c3e2f833 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.509560] env[61768]: DEBUG nova.objects.instance [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'flavor' on Instance uuid 8e0e13a7-eac5-4176-8536-2906d13c390e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.516157] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ad12a14-036c-4296-8342-05696d032f8c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.526760] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-809e24d4-4a22-40ed-ac37-4409361e3e0c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.541640] env[61768]: DEBUG nova.compute.provider_tree [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.601014] env[61768]: DEBUG nova.network.neutron [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updated VIF entry in instance network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 884.601454] env[61768]: DEBUG nova.network.neutron [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.666060] env[61768]: DEBUG nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.821891] env[61768]: DEBUG nova.network.neutron [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Successfully created port: 7084aac0-6dd6-411e-84e9-e4a2e353429c {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.902302] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.976784] env[61768]: DEBUG nova.network.neutron [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.021564] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e5815d39-618d-4797-8bac-386d87efa40e tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.380s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.045699] env[61768]: DEBUG nova.scheduler.client.report [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 885.109544] env[61768]: DEBUG oslo_concurrency.lockutils [req-bf991780-8a8a-41c5-86a6-933a55625d5c req-1bdcd7da-81e5-4a92-b34d-4db94d3d5f1f service nova] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.329124] env[61768]: DEBUG nova.network.neutron [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Updating instance_info_cache with network_info: [{"id": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "address": "fa:16:3e:a7:3a:6c", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc038ba6f-c7", "ovs_interfaceid": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.405635] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.529460] env[61768]: DEBUG nova.compute.manager [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.529460] env[61768]: DEBUG nova.compute.manager [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing instance network info cache due to event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.529460] env[61768]: DEBUG oslo_concurrency.lockutils [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.529460] env[61768]: DEBUG oslo_concurrency.lockutils [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.529460] env[61768]: DEBUG nova.network.neutron [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing network info cache for port 604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 885.558744] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.899s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.561985] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.680s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.563155] env[61768]: DEBUG nova.objects.instance [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lazy-loading 'resources' on Instance uuid aebe41ad-496d-4a53-b023-c8df2bca04f8 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 885.582484] env[61768]: INFO nova.scheduler.client.report [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted allocations for instance 3d96e47d-fc09-439e-bf0e-e60fc87ae958 [ 885.678922] env[61768]: DEBUG nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.687375] env[61768]: DEBUG nova.compute.manager [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.687743] env[61768]: DEBUG nova.compute.manager [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing instance network info cache due to event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.688145] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.713676] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.713676] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.713907] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.714013] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.714343] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.714686] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.714784] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.714977] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.715262] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.715521] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.715739] env[61768]: DEBUG nova.virt.hardware [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.716987] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63b15cca-2f68-4e0c-85d7-a20b17f2bb91 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.730071] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf73b51-c276-4166-8fb9-5fbe353343e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.832100] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.835018] env[61768]: DEBUG nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Instance network_info: |[{"id": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "address": "fa:16:3e:a7:3a:6c", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc038ba6f-c7", "ovs_interfaceid": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 885.835018] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:3a:6c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c038ba6f-c7aa-47fc-ad80-83bc55817296', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 885.843785] env[61768]: DEBUG oslo.service.loopingcall [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.844287] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 885.844656] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e5a930d-c6c6-464f-bded-c2e2233cccca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.874920] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 885.874920] env[61768]: value = "task-1229116" [ 885.874920] env[61768]: _type = "Task" [ 885.874920] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.888469] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229116, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.899818] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.912188] env[61768]: DEBUG nova.compute.manager [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Stashing vm_state: active {{(pid=61768) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 886.097153] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59e9d432-7beb-473d-8083-5d96d405cd8f tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "3d96e47d-fc09-439e-bf0e-e60fc87ae958" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.881s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.254610] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-ecd08348-b22f-49c5-b13f-b187506e38d6-20999117-9aaf-4291-9e72-f8304dfdf3f0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.254895] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-ecd08348-b22f-49c5-b13f-b187506e38d6-20999117-9aaf-4291-9e72-f8304dfdf3f0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.255344] env[61768]: DEBUG nova.objects.instance [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'flavor' on Instance uuid ecd08348-b22f-49c5-b13f-b187506e38d6 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.309657] env[61768]: DEBUG nova.network.neutron [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updated VIF entry in instance network info cache for port 604a3820-78ee-48e4-bb9e-af1273963ff8. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 886.310075] env[61768]: DEBUG nova.network.neutron [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.327835] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a60908-93ee-45f0-b064-c21e4c7c5abe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.336828] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487811ed-be3c-4a9b-a993-3fa1ad0c57a7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.368135] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad2b1ddb-5d2f-4472-809d-25bef3479ded {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.376131] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-614f8ec1-4a3b-4807-aef1-c833cab6a8a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.395049] env[61768]: DEBUG nova.compute.provider_tree [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.403061] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229116, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.404235] env[61768]: DEBUG nova.scheduler.client.report [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.415825] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.432153] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.755107] env[61768]: DEBUG nova.network.neutron [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Successfully updated port: 7084aac0-6dd6-411e-84e9-e4a2e353429c {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.812780] env[61768]: DEBUG oslo_concurrency.lockutils [req-b252183e-906f-4ca2-be99-4a98f63a6d32 req-3f39ad25-1ab7-4c14-95ec-fefdf4d481d4 service nova] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.813135] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.813414] env[61768]: DEBUG nova.network.neutron [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing network info cache for port 604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.862197] env[61768]: DEBUG nova.objects.instance [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'pci_requests' on Instance uuid ecd08348-b22f-49c5-b13f-b187506e38d6 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.892983] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229116, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.906054] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.910063] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.348s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.912402] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.759s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.912665] env[61768]: DEBUG nova.objects.instance [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'resources' on Instance uuid 12f92761-45b6-49c6-96bd-1c0fa05521c6 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 886.928197] env[61768]: INFO nova.scheduler.client.report [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted allocations for instance aebe41ad-496d-4a53-b023-c8df2bca04f8 [ 887.262345] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "refresh_cache-aaf205c5-dd11-4d1c-90bd-3ecd5a121227" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.262345] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "refresh_cache-aaf205c5-dd11-4d1c-90bd-3ecd5a121227" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.262848] env[61768]: DEBUG nova.network.neutron [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.318575] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "d54c94a6-0a53-47a8-b024-1a05439da837" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.318812] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "d54c94a6-0a53-47a8-b024-1a05439da837" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.364715] env[61768]: DEBUG nova.objects.base [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 887.364981] env[61768]: DEBUG nova.network.neutron [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 887.393905] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229116, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.406236] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.435336] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f2a25507-83de-4e4b-80d8-af99e543e6c3 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "aebe41ad-496d-4a53-b023-c8df2bca04f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.534s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.478728] env[61768]: DEBUG nova.policy [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 887.606548] env[61768]: DEBUG nova.compute.manager [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.606548] env[61768]: DEBUG nova.compute.manager [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing instance network info cache due to event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.606548] env[61768]: DEBUG oslo_concurrency.lockutils [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.606548] env[61768]: DEBUG oslo_concurrency.lockutils [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.606548] env[61768]: DEBUG nova.network.neutron [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 887.672078] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0c394da-c991-44ba-87d1-a405e0c47532 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.679361] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcb937d1-b9f8-40d4-a3d0-d54701b3a073 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.684504] env[61768]: DEBUG nova.network.neutron [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updated VIF entry in instance network info cache for port 604a3820-78ee-48e4-bb9e-af1273963ff8. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 887.684893] env[61768]: DEBUG nova.network.neutron [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.711343] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a25031b0-08cd-45f6-ad75-99b2b1bf2bed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.720762] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5299cd00-7ee9-4a3e-916d-a14ca0d88da9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.725970] env[61768]: DEBUG nova.compute.manager [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Received event network-vif-plugged-7084aac0-6dd6-411e-84e9-e4a2e353429c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.726238] env[61768]: DEBUG oslo_concurrency.lockutils [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] Acquiring lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.726524] env[61768]: DEBUG oslo_concurrency.lockutils [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.726631] env[61768]: DEBUG oslo_concurrency.lockutils [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.726807] env[61768]: DEBUG nova.compute.manager [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] No waiting events found dispatching network-vif-plugged-7084aac0-6dd6-411e-84e9-e4a2e353429c {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.726979] env[61768]: WARNING nova.compute.manager [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Received unexpected event network-vif-plugged-7084aac0-6dd6-411e-84e9-e4a2e353429c for instance with vm_state building and task_state spawning. [ 887.727163] env[61768]: DEBUG nova.compute.manager [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Received event network-changed-7084aac0-6dd6-411e-84e9-e4a2e353429c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.727556] env[61768]: DEBUG nova.compute.manager [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Refreshing instance network info cache due to event network-changed-7084aac0-6dd6-411e-84e9-e4a2e353429c. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.727556] env[61768]: DEBUG oslo_concurrency.lockutils [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] Acquiring lock "refresh_cache-aaf205c5-dd11-4d1c-90bd-3ecd5a121227" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.738594] env[61768]: DEBUG nova.compute.provider_tree [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 887.812712] env[61768]: DEBUG nova.network.neutron [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.821643] env[61768]: DEBUG nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 887.891901] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229116, 'name': CreateVM_Task, 'duration_secs': 1.870822} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.892164] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 887.892908] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.893221] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.893605] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.893912] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad440185-d5bb-4316-beca-5fb756223022 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.904754] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 887.904754] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5265b629-4285-9f93-73fd-341151fdd7de" [ 887.904754] env[61768]: _type = "Task" [ 887.904754] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.912440] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.918111] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5265b629-4285-9f93-73fd-341151fdd7de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.922741] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.922985] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.923198] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.923384] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.923559] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.926332] env[61768]: INFO nova.compute.manager [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Terminating instance [ 887.928088] env[61768]: DEBUG nova.compute.manager [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 887.928281] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 887.929070] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be8f218-fb3f-4aef-a5e6-9997c49f2bf8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.935949] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 887.936353] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fcb78620-eecf-4246-ae91-91027c506f10 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.943606] env[61768]: DEBUG oslo_vmware.api [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 887.943606] env[61768]: value = "task-1229117" [ 887.943606] env[61768]: _type = "Task" [ 887.943606] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.952841] env[61768]: DEBUG oslo_vmware.api [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229117, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.983137] env[61768]: DEBUG nova.network.neutron [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Updating instance_info_cache with network_info: [{"id": "7084aac0-6dd6-411e-84e9-e4a2e353429c", "address": "fa:16:3e:ce:24:95", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7084aac0-6d", "ovs_interfaceid": "7084aac0-6dd6-411e-84e9-e4a2e353429c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.211226] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.211645] env[61768]: DEBUG nova.compute.manager [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Received event network-vif-plugged-c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.211977] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Acquiring lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.212358] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.212642] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.212986] env[61768]: DEBUG nova.compute.manager [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] No waiting events found dispatching network-vif-plugged-c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 888.213259] env[61768]: WARNING nova.compute.manager [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Received unexpected event network-vif-plugged-c038ba6f-c7aa-47fc-ad80-83bc55817296 for instance with vm_state building and task_state spawning. [ 888.213470] env[61768]: DEBUG nova.compute.manager [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Received event network-changed-c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.213659] env[61768]: DEBUG nova.compute.manager [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Refreshing instance network info cache due to event network-changed-c038ba6f-c7aa-47fc-ad80-83bc55817296. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 888.213870] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Acquiring lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.214027] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Acquired lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.214199] env[61768]: DEBUG nova.network.neutron [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Refreshing network info cache for port c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.241935] env[61768]: DEBUG nova.scheduler.client.report [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 888.344569] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.349523] env[61768]: DEBUG nova.network.neutron [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updated VIF entry in instance network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.349933] env[61768]: DEBUG nova.network.neutron [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.407673] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229114, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.583979} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.410853] env[61768]: INFO nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0/OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0.vmdk to [datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3/1e83b86d-089f-4b79-940c-ef76d051d4f3.vmdk. [ 888.411075] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Cleaning up location [datastore2] OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 888.411256] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_ba6eee77-ba60-41fa-924f-f7915ff3b5b0 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.411522] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c8d478a-caf4-425f-9cf6-9f636b52e2c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.418349] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5265b629-4285-9f93-73fd-341151fdd7de, 'name': SearchDatastore_Task, 'duration_secs': 0.052212} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.419734] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.419963] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.420248] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.420416] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.420606] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.420954] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 888.420954] env[61768]: value = "task-1229118" [ 888.420954] env[61768]: _type = "Task" [ 888.420954] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.421162] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34f50e3e-ffeb-4924-bb7e-1e2bb1183360 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.431364] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229118, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.434482] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.434698] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 888.435487] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-647c8ca0-0fa9-4a90-91c3-180c3bddb09c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.440537] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 888.440537] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528f618b-ae6b-5f04-4cd8-87fd93b70f4d" [ 888.440537] env[61768]: _type = "Task" [ 888.440537] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.449629] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528f618b-ae6b-5f04-4cd8-87fd93b70f4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.454119] env[61768]: DEBUG oslo_vmware.api [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229117, 'name': PowerOffVM_Task, 'duration_secs': 0.298504} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.454429] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 888.454582] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 888.454823] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8ff65f21-571d-40d1-beb0-13b9c4378252 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.486077] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "refresh_cache-aaf205c5-dd11-4d1c-90bd-3ecd5a121227" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.486423] env[61768]: DEBUG nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Instance network_info: |[{"id": "7084aac0-6dd6-411e-84e9-e4a2e353429c", "address": "fa:16:3e:ce:24:95", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7084aac0-6d", "ovs_interfaceid": "7084aac0-6dd6-411e-84e9-e4a2e353429c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 888.486833] env[61768]: DEBUG oslo_concurrency.lockutils [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] Acquired lock "refresh_cache-aaf205c5-dd11-4d1c-90bd-3ecd5a121227" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.487077] env[61768]: DEBUG nova.network.neutron [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Refreshing network info cache for port 7084aac0-6dd6-411e-84e9-e4a2e353429c {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 888.488410] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ce:24:95', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '680cb499-2a47-482b-af0d-112016ac0e17', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7084aac0-6dd6-411e-84e9-e4a2e353429c', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 888.496776] env[61768]: DEBUG oslo.service.loopingcall [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 888.500071] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 888.501050] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20cae1c5-9de8-44aa-bda5-3f89d31c9ab6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.519147] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 888.519432] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 888.519693] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleting the datastore file [datastore2] 23cc444d-ff15-441b-8a0c-2ded49fd3cf0 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 888.520181] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96e63e2c-306d-4d5e-a2cf-e6c6b75563db {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.524985] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 888.524985] env[61768]: value = "task-1229120" [ 888.524985] env[61768]: _type = "Task" [ 888.524985] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.529616] env[61768]: DEBUG oslo_vmware.api [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 888.529616] env[61768]: value = "task-1229121" [ 888.529616] env[61768]: _type = "Task" [ 888.529616] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.536070] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229120, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.540920] env[61768]: DEBUG oslo_vmware.api [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229121, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.716090] env[61768]: DEBUG nova.network.neutron [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Updated VIF entry in instance network info cache for port 7084aac0-6dd6-411e-84e9-e4a2e353429c. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.716590] env[61768]: DEBUG nova.network.neutron [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Updating instance_info_cache with network_info: [{"id": "7084aac0-6dd6-411e-84e9-e4a2e353429c", "address": "fa:16:3e:ce:24:95", "network": {"id": "a10562ef-ef71-4e5a-8f89-4a47495e029f", "bridge": "br-int", "label": "tempest-ImagesTestJSON-1047223830-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "180da0bcd5954eb2bc89a20cf8f42bb5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "680cb499-2a47-482b-af0d-112016ac0e17", "external-id": "nsx-vlan-transportzone-644", "segmentation_id": 644, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7084aac0-6d", "ovs_interfaceid": "7084aac0-6dd6-411e-84e9-e4a2e353429c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.747222] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.835s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.752284] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.320s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.768514] env[61768]: INFO nova.scheduler.client.report [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted allocations for instance 12f92761-45b6-49c6-96bd-1c0fa05521c6 [ 888.855264] env[61768]: DEBUG oslo_concurrency.lockutils [req-b7135dc7-b923-4fc8-86b3-0c3e423c9cb1 req-c991a923-f421-4494-aefa-b13509dee0bc service nova] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.934952] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229118, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140053} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.935235] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 888.935424] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3/1e83b86d-089f-4b79-940c-ef76d051d4f3.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.935681] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3/1e83b86d-089f-4b79-940c-ef76d051d4f3.vmdk to [datastore2] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 888.935947] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe137b99-9ada-488b-a445-9d41a12a7615 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.941949] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 888.941949] env[61768]: value = "task-1229122" [ 888.941949] env[61768]: _type = "Task" [ 888.941949] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.954653] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528f618b-ae6b-5f04-4cd8-87fd93b70f4d, 'name': SearchDatastore_Task, 'duration_secs': 0.045885} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.962695] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229122, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.962695] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a28d8155-990b-4371-9bed-e8ca09083d2b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.966206] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 888.966206] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525013dc-5302-ee83-c220-ee6e34fb4dc9" [ 888.966206] env[61768]: _type = "Task" [ 888.966206] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.970213] env[61768]: DEBUG nova.network.neutron [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Updated VIF entry in instance network info cache for port c038ba6f-c7aa-47fc-ad80-83bc55817296. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 888.970592] env[61768]: DEBUG nova.network.neutron [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Updating instance_info_cache with network_info: [{"id": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "address": "fa:16:3e:a7:3a:6c", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc038ba6f-c7", "ovs_interfaceid": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.977535] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525013dc-5302-ee83-c220-ee6e34fb4dc9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.039023] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229120, 'name': CreateVM_Task, 'duration_secs': 0.414796} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.039023] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 889.039023] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.039657] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.040208] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 889.043869] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47477eec-32c8-4452-8023-59c8573e3660 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.046288] env[61768]: DEBUG oslo_vmware.api [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229121, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.386352} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.047145] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 889.047492] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 889.048689] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 889.048689] env[61768]: INFO nova.compute.manager [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Took 1.12 seconds to destroy the instance on the hypervisor. [ 889.048689] env[61768]: DEBUG oslo.service.loopingcall [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.049216] env[61768]: DEBUG nova.compute.manager [-] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 889.049482] env[61768]: DEBUG nova.network.neutron [-] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 889.055111] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 889.055111] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5297e1dd-1631-9b41-0513-31d6002df057" [ 889.055111] env[61768]: _type = "Task" [ 889.055111] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.061675] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5297e1dd-1631-9b41-0513-31d6002df057, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.223153] env[61768]: DEBUG oslo_concurrency.lockutils [req-7a07c482-bba1-42cf-ac1e-d62d225fe39b req-e33e04e5-ae2c-4ee4-80bd-530f79bdf11b service nova] Releasing lock "refresh_cache-aaf205c5-dd11-4d1c-90bd-3ecd5a121227" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.257988] env[61768]: INFO nova.compute.claims [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.275777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-13f94b25-8ac9-42a6-bd64-2944f33163cd tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "12f92761-45b6-49c6-96bd-1c0fa05521c6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.091s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.308926] env[61768]: DEBUG nova.network.neutron [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Successfully updated port: 20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 889.455655] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229122, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.473359] env[61768]: DEBUG oslo_concurrency.lockutils [req-bb777977-df8b-4435-abd2-9c90a71ae14d req-b5473a83-1d3e-4155-8c18-18a5e23871d7 service nova] Releasing lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.482910] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525013dc-5302-ee83-c220-ee6e34fb4dc9, 'name': SearchDatastore_Task, 'duration_secs': 0.044407} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.483013] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.483231] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015/4cc5b1fa-55e3-4b2d-b10d-8da7b3356015.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 889.483540] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c5726e05-8aeb-4d98-9da6-3485e7d017d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.490706] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 889.490706] env[61768]: value = "task-1229123" [ 889.490706] env[61768]: _type = "Task" [ 889.490706] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.499875] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.567087] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5297e1dd-1631-9b41-0513-31d6002df057, 'name': SearchDatastore_Task, 'duration_secs': 0.027073} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.567498] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.567745] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 889.568028] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.568198] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.568394] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 889.568677] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03904de8-5841-47fd-830e-3557634d962f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.587183] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 889.587462] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 889.588357] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-570704c1-6029-4ed6-8b53-912916b75340 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.596238] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 889.596238] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52664849-9fa0-46ef-5408-f52a147d64e8" [ 889.596238] env[61768]: _type = "Task" [ 889.596238] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.605784] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52664849-9fa0-46ef-5408-f52a147d64e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.764222] env[61768]: INFO nova.compute.resource_tracker [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating resource usage from migration 68c66c79-2d4b-4e76-a092-90d11ed31f0b [ 889.811690] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.811909] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.812118] env[61768]: DEBUG nova.network.neutron [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.916381] env[61768]: DEBUG nova.network.neutron [-] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.960152] env[61768]: DEBUG nova.compute.manager [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-vif-plugged-20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.961356] env[61768]: DEBUG oslo_concurrency.lockutils [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] Acquiring lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.961620] env[61768]: DEBUG oslo_concurrency.lockutils [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.961814] env[61768]: DEBUG oslo_concurrency.lockutils [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.962009] env[61768]: DEBUG nova.compute.manager [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] No waiting events found dispatching network-vif-plugged-20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 889.962203] env[61768]: WARNING nova.compute.manager [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received unexpected event network-vif-plugged-20999117-9aaf-4291-9e72-f8304dfdf3f0 for instance with vm_state active and task_state None. [ 889.962533] env[61768]: DEBUG nova.compute.manager [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-changed-20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.962533] env[61768]: DEBUG nova.compute.manager [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing instance network info cache due to event network-changed-20999117-9aaf-4291-9e72-f8304dfdf3f0. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.962721] env[61768]: DEBUG oslo_concurrency.lockutils [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.970155] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229122, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.001948] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.003970] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d1e4c4-2860-420a-a72a-dfd6562e4600 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.013598] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954912ea-3791-4e53-914f-08aec8d0f484 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.049533] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed59256e-ebc0-4616-87a4-92e264053041 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.058516] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2491b31-48df-4271-b36e-4e0b77d76ab4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.075209] env[61768]: DEBUG nova.compute.provider_tree [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.107238] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52664849-9fa0-46ef-5408-f52a147d64e8, 'name': SearchDatastore_Task, 'duration_secs': 0.08069} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.108144] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8179c422-ac95-4518-b129-b771376a0655 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.116308] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 890.116308] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b12170-6f1a-fccf-d92b-01616c5604b7" [ 890.116308] env[61768]: _type = "Task" [ 890.116308] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.125929] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b12170-6f1a-fccf-d92b-01616c5604b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.353504] env[61768]: WARNING nova.network.neutron [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] 2fbe88f9-1bca-4347-914d-a62dc6bebde8 already exists in list: networks containing: ['2fbe88f9-1bca-4347-914d-a62dc6bebde8']. ignoring it [ 890.396011] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "739faae0-5811-4f88-b56b-1350c0f7b8be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.396316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.419411] env[61768]: INFO nova.compute.manager [-] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Took 1.37 seconds to deallocate network for instance. [ 890.457602] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229122, 'name': CopyVirtualDisk_Task} progress is 54%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.504156] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229123, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.578882] env[61768]: DEBUG nova.scheduler.client.report [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.629088] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52b12170-6f1a-fccf-d92b-01616c5604b7, 'name': SearchDatastore_Task, 'duration_secs': 0.088601} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.629581] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.629685] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] aaf205c5-dd11-4d1c-90bd-3ecd5a121227/aaf205c5-dd11-4d1c-90bd-3ecd5a121227.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 890.629962] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-901075fb-49db-4c83-b22c-be0623f74362 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.641421] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 890.641421] env[61768]: value = "task-1229125" [ 890.641421] env[61768]: _type = "Task" [ 890.641421] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.651702] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229125, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.711174] env[61768]: DEBUG nova.network.neutron [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "address": "fa:16:3e:ad:82:00", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20999117-9a", "ovs_interfaceid": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.898823] env[61768]: DEBUG nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.926603] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.958416] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229122, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.005189] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229123, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.086531] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.334s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.086836] env[61768]: INFO nova.compute.manager [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Migrating [ 891.093641] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.749s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.095186] env[61768]: INFO nova.compute.claims [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.154645] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229125, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.214701] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.215515] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.215680] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.216785] env[61768]: DEBUG oslo_concurrency.lockutils [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.217057] env[61768]: DEBUG nova.network.neutron [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing network info cache for port 20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.219251] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549cdef1-ddfa-4de5-8636-18f2b864613d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.241140] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.241436] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.241610] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.241799] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.241958] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.242123] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.242341] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.242510] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.242684] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.242941] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.243237] env[61768]: DEBUG nova.virt.hardware [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.249634] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Reconfiguring VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 891.250344] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98779a35-5f43-4f78-ae46-3c996e4a56ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.271389] env[61768]: DEBUG oslo_vmware.api [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 891.271389] env[61768]: value = "task-1229126" [ 891.271389] env[61768]: _type = "Task" [ 891.271389] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.281955] env[61768]: DEBUG oslo_vmware.api [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229126, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.427416] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.459724] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229122, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.503924] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229123, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.608631] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.608843] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.608994] env[61768]: DEBUG nova.network.neutron [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.655246] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229125, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.781562] env[61768]: DEBUG oslo_vmware.api [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229126, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.943794] env[61768]: DEBUG nova.network.neutron [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updated VIF entry in instance network info cache for port 20999117-9aaf-4291-9e72-f8304dfdf3f0. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 891.944478] env[61768]: DEBUG nova.network.neutron [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "address": "fa:16:3e:ad:82:00", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20999117-9a", "ovs_interfaceid": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.959401] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229122, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.683205} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.960376] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/1e83b86d-089f-4b79-940c-ef76d051d4f3/1e83b86d-089f-4b79-940c-ef76d051d4f3.vmdk to [datastore2] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 891.961233] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77f6789-057a-4c5d-b868-2147931d9109 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.987637] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk or device None with type streamOptimized {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 891.988351] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8efd2cd0-074e-40e8-b007-bd3d0b0dc181 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.013284] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229123, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.280616} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.014876] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015/4cc5b1fa-55e3-4b2d-b10d-8da7b3356015.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 892.015141] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.015484] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 892.015484] env[61768]: value = "task-1229128" [ 892.015484] env[61768]: _type = "Task" [ 892.015484] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.015685] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64a9d76e-c731-418e-a72d-ffe509a06308 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.027382] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229128, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.029051] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 892.029051] env[61768]: value = "task-1229129" [ 892.029051] env[61768]: _type = "Task" [ 892.029051] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.041320] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229129, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.157600] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229125, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.507694} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.157873] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] aaf205c5-dd11-4d1c-90bd-3ecd5a121227/aaf205c5-dd11-4d1c-90bd-3ecd5a121227.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 892.158115] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.158383] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13cccab9-4f3d-44a7-a9bf-306617f6a403 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.165843] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 892.165843] env[61768]: value = "task-1229130" [ 892.165843] env[61768]: _type = "Task" [ 892.165843] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.173631] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.281781] env[61768]: DEBUG oslo_vmware.api [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229126, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.333138] env[61768]: DEBUG nova.network.neutron [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance_info_cache with network_info: [{"id": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "address": "fa:16:3e:2b:a2:c7", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4c6981-f4", "ovs_interfaceid": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.351770] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7506fbc3-6165-428e-a810-c3887ce7f297 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.361145] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8db18b7e-5219-434f-8109-6d502e2dcd64 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.394274] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3198186a-b46f-4c6e-aa6a-d575b1635a0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.402276] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8612dbaf-0559-4f9a-ac53-beaf127ed7d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.416630] env[61768]: DEBUG nova.compute.provider_tree [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.447791] env[61768]: DEBUG oslo_concurrency.lockutils [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.447965] env[61768]: DEBUG nova.compute.manager [req-170a6e78-0253-465a-bc00-cca95279f196 req-c891a1a0-75b9-47bb-8e4d-16f952252ab6 service nova] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Received event network-vif-deleted-5405ba6b-c888-48a0-9af2-4024e9af563e {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.528213] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229128, 'name': ReconfigVM_Task, 'duration_secs': 0.341012} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.528532] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfigured VM instance instance-0000003e to attach disk [datastore2] 62cddff5-d499-4ef3-869d-3cdb0328640d/62cddff5-d499-4ef3-869d-3cdb0328640d.vmdk or device None with type streamOptimized {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 892.529941] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'image_type': None, 'device_type': 'disk', 'device_name': '/dev/sda', 'guest_format': None, 'encryption_secret_uuid': None, 'encryption_format': None, 'size': 0, 'encryption_options': None, 'disk_bus': None, 'boot_index': 0, 'encrypted': False, 'image_id': '0a07405a-3ac2-4d7f-bd39-a1b659ab2354'}], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': False, 'device_type': None, 'guest_format': None, 'mount_device': '/dev/sdb', 'attachment_id': 'c8ecb62d-95ff-4d4c-a115-bf03b0d110f5', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265482', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'name': 'volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '62cddff5-d499-4ef3-869d-3cdb0328640d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'serial': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae'}, 'disk_bus': None, 'boot_index': None, 'volume_type': None}], 'swap': None} {{(pid=61768) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 892.530175] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 892.530467] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265482', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'name': 'volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '62cddff5-d499-4ef3-869d-3cdb0328640d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'serial': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 892.531220] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9afac08-40fc-43a3-8c4e-ebe62862220a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.541598] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229129, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085139} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.555052] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.555587] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eafd9448-51e1-41b9-9a0d-a46241300bc7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.558951] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466dae4c-9d09-4d28-bb21-371b9d233759 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.597176] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfiguring VM instance instance-0000003e to attach disk [datastore2] volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae/volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.620541] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6780e88-7ff1-44e9-ad30-c0eca7b44385 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.646796] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015/4cc5b1fa-55e3-4b2d-b10d-8da7b3356015.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.647163] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e052801-2d94-43ff-a418-07adb2d647db {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.669029] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 892.669029] env[61768]: value = "task-1229132" [ 892.669029] env[61768]: _type = "Task" [ 892.669029] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.675201] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 892.675201] env[61768]: value = "task-1229131" [ 892.675201] env[61768]: _type = "Task" [ 892.675201] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.682469] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110182} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.683017] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.683814] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa64570d-ca96-4251-8ce9-6fbe5d24b42b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.692871] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229131, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.693614] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229132, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.727680] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] aaf205c5-dd11-4d1c-90bd-3ecd5a121227/aaf205c5-dd11-4d1c-90bd-3ecd5a121227.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.728172] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0da0358e-cad9-4b16-82cb-0d606f67d63a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.758936] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 892.758936] env[61768]: value = "task-1229133" [ 892.758936] env[61768]: _type = "Task" [ 892.758936] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.768067] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229133, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.781858] env[61768]: DEBUG oslo_vmware.api [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229126, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.838714] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.920738] env[61768]: DEBUG nova.scheduler.client.report [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.185768] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229132, 'name': ReconfigVM_Task, 'duration_secs': 0.372021} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.188631] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015/4cc5b1fa-55e3-4b2d-b10d-8da7b3356015.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.189301] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229131, 'name': ReconfigVM_Task, 'duration_secs': 0.306818} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.189541] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4896e73d-8c87-4107-a03f-f6d1d2f94e1a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.191160] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfigured VM instance instance-0000003e to attach disk [datastore2] volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae/volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.195702] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01fcc4ae-1329-47cb-91cb-a10c964e0bca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.211394] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 893.211394] env[61768]: value = "task-1229134" [ 893.211394] env[61768]: _type = "Task" [ 893.211394] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.212320] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 893.212320] env[61768]: value = "task-1229135" [ 893.212320] env[61768]: _type = "Task" [ 893.212320] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.222671] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229134, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.225612] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229135, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.268210] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229133, 'name': ReconfigVM_Task, 'duration_secs': 0.289559} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.268554] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Reconfigured VM instance instance-00000058 to attach disk [datastore2] aaf205c5-dd11-4d1c-90bd-3ecd5a121227/aaf205c5-dd11-4d1c-90bd-3ecd5a121227.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.269261] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ef75f04-e7c9-4b30-b66f-2bc40fdd1aeb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.278503] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 893.278503] env[61768]: value = "task-1229136" [ 893.278503] env[61768]: _type = "Task" [ 893.278503] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.286779] env[61768]: DEBUG oslo_vmware.api [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229126, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.291631] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229136, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.426373] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.333s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.426921] env[61768]: DEBUG nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.430236] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.504s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.430471] env[61768]: DEBUG nova.objects.instance [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lazy-loading 'resources' on Instance uuid 23cc444d-ff15-441b-8a0c-2ded49fd3cf0 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.483358] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.483748] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.725279] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229134, 'name': Rename_Task, 'duration_secs': 0.157667} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.728256] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.728534] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229135, 'name': ReconfigVM_Task, 'duration_secs': 0.149073} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.728737] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bcabf8d-7e62-49eb-9a42-3931fb042717 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.730262] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265482', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'name': 'volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '62cddff5-d499-4ef3-869d-3cdb0328640d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'serial': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 893.730814] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ce5fbdc8-3cb4-4676-a5c9-ba036a0a3ddd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.737261] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 893.737261] env[61768]: value = "task-1229137" [ 893.737261] env[61768]: _type = "Task" [ 893.737261] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.738531] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 893.738531] env[61768]: value = "task-1229138" [ 893.738531] env[61768]: _type = "Task" [ 893.738531] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.749382] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229137, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.752433] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229138, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.787062] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229136, 'name': Rename_Task, 'duration_secs': 0.160475} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.790198] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 893.790457] env[61768]: DEBUG oslo_vmware.api [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229126, 'name': ReconfigVM_Task, 'duration_secs': 2.203672} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.790666] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-771d20e1-d119-45d9-8b46-458071b870dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.792326] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.792539] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Reconfigured VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 893.802020] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 893.802020] env[61768]: value = "task-1229139" [ 893.802020] env[61768]: _type = "Task" [ 893.802020] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.810185] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229139, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.934193] env[61768]: DEBUG nova.compute.utils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.938710] env[61768]: DEBUG nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.943135] env[61768]: DEBUG nova.network.neutron [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 893.984346] env[61768]: DEBUG nova.policy [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2280e9e6bde740c3aca30ebf70c7f73d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc444d04ff864300a45a55a3c67507e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.986643] env[61768]: INFO nova.compute.manager [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Detaching volume a38ffb49-6325-4242-819b-ba4b116389d5 [ 894.024564] env[61768]: INFO nova.virt.block_device [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Attempting to driver detach volume a38ffb49-6325-4242-819b-ba4b116389d5 from mountpoint /dev/sdb [ 894.024863] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 894.025074] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265474', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'name': 'volume-a38ffb49-6325-4242-819b-ba4b116389d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7697ecda-ef30-447d-a92a-3f5cb3cc9118', 'attached_at': '', 'detached_at': '', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'serial': 'a38ffb49-6325-4242-819b-ba4b116389d5'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 894.025962] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f86186a-e672-4ccb-bf3e-4014899ca44a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.055072] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b832731b-bf37-462a-b7b2-3292f9908bb7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.062341] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002ccdbd-068f-4528-833e-015a5c6f0934 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.086062] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e13eb20-1054-4d41-bffa-2e8ab509a31b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.101342] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] The volume has not been displaced from its original location: [datastore2] volume-a38ffb49-6325-4242-819b-ba4b116389d5/volume-a38ffb49-6325-4242-819b-ba4b116389d5.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 894.106574] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Reconfiguring VM instance instance-00000043 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 894.109459] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5c3907d1-8600-4bd1-8f66-bbf329b271fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.128690] env[61768]: DEBUG oslo_vmware.api [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 894.128690] env[61768]: value = "task-1229141" [ 894.128690] env[61768]: _type = "Task" [ 894.128690] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.137013] env[61768]: DEBUG oslo_vmware.api [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229141, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.219216] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819d36c4-4d0d-46c7-8e58-1b1cd835d68f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.227977] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4fab26-5e9b-4051-8472-e45924b64f09 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.268363] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15dc4958-9c90-4296-b1ea-acf839a08cf2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.276877] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229137, 'name': Rename_Task, 'duration_secs': 0.184939} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.282674] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 894.283136] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229138, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.283507] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a3b1b22d-48b6-4ca2-82be-ebc041a397ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.286372] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47abf9c2-3feb-4a6c-ae05-16ea94d28a47 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.297383] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b86d3b5d-bfff-4b29-9bfa-d8588baf4482 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-ecd08348-b22f-49c5-b13f-b187506e38d6-20999117-9aaf-4291-9e72-f8304dfdf3f0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.042s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.307951] env[61768]: DEBUG nova.compute.provider_tree [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.309715] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 894.309715] env[61768]: value = "task-1229142" [ 894.309715] env[61768]: _type = "Task" [ 894.309715] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.326158] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229139, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.330562] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229142, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.336477] env[61768]: DEBUG nova.network.neutron [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Successfully created port: 7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.356121] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5413c18-a635-4f25-bcb9-e069500847d2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.379081] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance 'e2b7cb22-475b-4b1b-b30e-6061e888691e' progress to 0 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 894.440107] env[61768]: DEBUG nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.638631] env[61768]: DEBUG oslo_vmware.api [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229141, 'name': ReconfigVM_Task, 'duration_secs': 0.265486} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.639044] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Reconfigured VM instance instance-00000043 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 894.643850] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12543645-d0f2-4320-93d9-fe831e2a5fd3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.659451] env[61768]: DEBUG oslo_vmware.api [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 894.659451] env[61768]: value = "task-1229143" [ 894.659451] env[61768]: _type = "Task" [ 894.659451] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.667198] env[61768]: DEBUG oslo_vmware.api [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229143, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.771792] env[61768]: DEBUG oslo_vmware.api [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229138, 'name': PowerOnVM_Task, 'duration_secs': 0.544992} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.772272] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 894.772447] env[61768]: INFO nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Took 11.57 seconds to spawn the instance on the hypervisor. [ 894.772678] env[61768]: DEBUG nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.773620] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea92192-c989-4960-bd77-541fab9dbbb3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.815558] env[61768]: DEBUG nova.scheduler.client.report [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.826821] env[61768]: DEBUG oslo_vmware.api [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229139, 'name': PowerOnVM_Task, 'duration_secs': 0.604799} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.826821] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 894.827116] env[61768]: INFO nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Took 9.15 seconds to spawn the instance on the hypervisor. [ 894.827405] env[61768]: DEBUG nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 894.829551] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0fed4d-2769-479d-ac7d-b22a2fad8910 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.838170] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229142, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.886055] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.886777] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d991fa1f-b854-485e-a2ba-1b1b1c2310f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.894093] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 894.894093] env[61768]: value = "task-1229144" [ 894.894093] env[61768]: _type = "Task" [ 894.894093] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.905260] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229144, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.170878] env[61768]: DEBUG oslo_vmware.api [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229143, 'name': ReconfigVM_Task, 'duration_secs': 0.144306} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.171264] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265474', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'name': 'volume-a38ffb49-6325-4242-819b-ba4b116389d5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '7697ecda-ef30-447d-a92a-3f5cb3cc9118', 'attached_at': '', 'detached_at': '', 'volume_id': 'a38ffb49-6325-4242-819b-ba4b116389d5', 'serial': 'a38ffb49-6325-4242-819b-ba4b116389d5'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 895.293666] env[61768]: INFO nova.compute.manager [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Took 27.08 seconds to build instance. [ 895.324540] env[61768]: DEBUG oslo_vmware.api [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229142, 'name': PowerOnVM_Task, 'duration_secs': 0.675732} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.324832] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 895.327390] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.897s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.329696] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.902s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.330885] env[61768]: INFO nova.compute.claims [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.351691] env[61768]: INFO nova.scheduler.client.report [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted allocations for instance 23cc444d-ff15-441b-8a0c-2ded49fd3cf0 [ 895.359755] env[61768]: INFO nova.compute.manager [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Took 25.59 seconds to build instance. [ 895.405849] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229144, 'name': PowerOffVM_Task, 'duration_secs': 0.36295} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.406247] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 895.406761] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance 'e2b7cb22-475b-4b1b-b30e-6061e888691e' progress to 17 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 895.455435] env[61768]: DEBUG nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.467829] env[61768]: DEBUG nova.compute.manager [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.469058] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b11bc899-dd07-460b-b9e7-d297009144e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.487632] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.487940] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.488152] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.488379] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.489099] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.489099] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.489099] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.489278] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.489402] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.489605] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.489788] env[61768]: DEBUG nova.virt.hardware [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.490654] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da6fc6b-200f-43c8-99b8-eef5bcbd8566 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.499453] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed7d339-49f8-483b-b5bc-d42c95d451cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.748941] env[61768]: DEBUG nova.objects.instance [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 895.795657] env[61768]: DEBUG oslo_concurrency.lockutils [None req-40707e7a-f4fb-48b8-9557-9d91a534ae87 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.593s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.860545] env[61768]: DEBUG oslo_concurrency.lockutils [None req-cef107c8-c572-4448-b00f-226ee5efd0a1 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "23cc444d-ff15-441b-8a0c-2ded49fd3cf0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.937s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.865021] env[61768]: DEBUG oslo_concurrency.lockutils [None req-305cb31f-a578-493c-9cbb-b71b3eb2063c tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.109s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.912946] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.913231] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.913400] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.913594] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.913751] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.913908] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.914385] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.914599] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.914895] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.915117] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.915311] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.921787] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e0cdd79-fce7-4371-88cd-2a4f62b64298 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.938812] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 895.938812] env[61768]: value = "task-1229145" [ 895.938812] env[61768]: _type = "Task" [ 895.938812] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.948154] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229145, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.986761] env[61768]: DEBUG oslo_concurrency.lockutils [None req-61a19740-3d79-40ec-bb83-06894eba69c5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 37.453s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.007289] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.007628] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.007844] env[61768]: INFO nova.compute.manager [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Shelving [ 896.074515] env[61768]: DEBUG nova.compute.manager [req-17f35d26-c693-453c-930a-67fd44c02b54 req-6948a98c-1287-4753-ae60-17f0bea0ce66 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Received event network-vif-plugged-7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.074755] env[61768]: DEBUG oslo_concurrency.lockutils [req-17f35d26-c693-453c-930a-67fd44c02b54 req-6948a98c-1287-4753-ae60-17f0bea0ce66 service nova] Acquiring lock "d54c94a6-0a53-47a8-b024-1a05439da837-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.075156] env[61768]: DEBUG oslo_concurrency.lockutils [req-17f35d26-c693-453c-930a-67fd44c02b54 req-6948a98c-1287-4753-ae60-17f0bea0ce66 service nova] Lock "d54c94a6-0a53-47a8-b024-1a05439da837-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.075518] env[61768]: DEBUG oslo_concurrency.lockutils [req-17f35d26-c693-453c-930a-67fd44c02b54 req-6948a98c-1287-4753-ae60-17f0bea0ce66 service nova] Lock "d54c94a6-0a53-47a8-b024-1a05439da837-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.075844] env[61768]: DEBUG nova.compute.manager [req-17f35d26-c693-453c-930a-67fd44c02b54 req-6948a98c-1287-4753-ae60-17f0bea0ce66 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] No waiting events found dispatching network-vif-plugged-7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.076065] env[61768]: WARNING nova.compute.manager [req-17f35d26-c693-453c-930a-67fd44c02b54 req-6948a98c-1287-4753-ae60-17f0bea0ce66 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Received unexpected event network-vif-plugged-7660d6d5-91ae-41e4-ba9f-1c09038a38c1 for instance with vm_state building and task_state spawning. [ 896.087412] env[61768]: DEBUG nova.network.neutron [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Successfully updated port: 7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.281618] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-ecd08348-b22f-49c5-b13f-b187506e38d6-20999117-9aaf-4291-9e72-f8304dfdf3f0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.282043] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-ecd08348-b22f-49c5-b13f-b187506e38d6-20999117-9aaf-4291-9e72-f8304dfdf3f0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.450860] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229145, 'name': ReconfigVM_Task, 'duration_secs': 0.176233} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.451293] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance 'e2b7cb22-475b-4b1b-b30e-6061e888691e' progress to 33 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 896.517497] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 896.521043] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-990ddd3f-300b-45f3-b8dd-b53b0d109bb2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.529321] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 896.529321] env[61768]: value = "task-1229147" [ 896.529321] env[61768]: _type = "Task" [ 896.529321] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.548554] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229147, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.573708] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c61e893-d202-4da8-8a5c-8ac4b830d53a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.581485] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8507e778-23ab-4d49-979d-2e46434bee74 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.589855] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.589995] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.590161] env[61768]: DEBUG nova.network.neutron [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 896.622160] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959aad87-f387-462b-8f99-613341e5fba4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.630080] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae9191a-be16-4eb4-9810-e72885f93faa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.643494] env[61768]: DEBUG nova.compute.provider_tree [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.648758] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.648758] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.648942] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.649082] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.649257] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.651358] env[61768]: INFO nova.compute.manager [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Terminating instance [ 896.653167] env[61768]: DEBUG nova.compute.manager [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 896.653367] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 896.654111] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59559e53-2339-4001-ba85-6052a696d80e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.660952] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 896.661381] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-46f37b1a-02ef-425b-92d5-f05647f24bc4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.667730] env[61768]: DEBUG oslo_vmware.api [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 896.667730] env[61768]: value = "task-1229148" [ 896.667730] env[61768]: _type = "Task" [ 896.667730] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.675231] env[61768]: DEBUG oslo_vmware.api [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.757065] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4104bd96-cece-4f47-817c-506e1fad3805 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.273s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.774774] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.775244] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.775451] env[61768]: DEBUG nova.compute.manager [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.776324] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-341aa598-7d4c-44eb-98f5-7b43aff784ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.783090] env[61768]: DEBUG nova.compute.manager [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61768) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 896.783660] env[61768]: DEBUG nova.objects.instance [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 896.785483] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.785483] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.786251] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce59b289-efaa-43f0-a31e-1af5dcc5fe84 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.807088] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5835367c-4e53-4f9a-a80f-aef5bd371ab7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.837419] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Reconfiguring VM to detach interface {{(pid=61768) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 896.838123] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-245d746d-0072-4864-8173-9524c958db96 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.857270] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 896.857270] env[61768]: value = "task-1229149" [ 896.857270] env[61768]: _type = "Task" [ 896.857270] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.866683] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.958672] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 896.958988] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 896.959156] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 896.959354] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 896.959685] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 896.959899] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 896.960181] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 896.960371] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 896.960583] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 896.960767] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 896.960951] env[61768]: DEBUG nova.virt.hardware [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 896.966542] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 896.967179] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-323308e5-e375-4482-a19d-1cacdd87acf2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.987088] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 896.987088] env[61768]: value = "task-1229150" [ 896.987088] env[61768]: _type = "Task" [ 896.987088] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.994986] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229150, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.039400] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229147, 'name': PowerOffVM_Task, 'duration_secs': 0.233261} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.039802] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 897.040616] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088663b0-e53f-4a4d-bfdd-c80e6302073c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.060683] env[61768]: DEBUG nova.compute.manager [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.061716] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5809ea50-f833-498b-816c-ecd21554e15b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.065266] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4cbb336-2ed4-40e5-8146-53b3f3c3aba9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.124313] env[61768]: DEBUG nova.network.neutron [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 897.146332] env[61768]: DEBUG nova.scheduler.client.report [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.176170] env[61768]: DEBUG oslo_vmware.api [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229148, 'name': PowerOffVM_Task, 'duration_secs': 0.18127} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.176443] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 897.176619] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 897.176909] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5dc977cd-e84c-4d3b-aa99-41ca5363f031 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.233451] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 897.233791] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 897.233994] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleting the datastore file [datastore1] 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 897.234281] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96f36c46-f71e-4ebe-ad51-478f56e003fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.240022] env[61768]: DEBUG oslo_vmware.api [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 897.240022] env[61768]: value = "task-1229152" [ 897.240022] env[61768]: _type = "Task" [ 897.240022] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.247673] env[61768]: DEBUG oslo_vmware.api [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229152, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.260796] env[61768]: DEBUG nova.network.neutron [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Updating instance_info_cache with network_info: [{"id": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "address": "fa:16:3e:52:a7:22", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7660d6d5-91", "ovs_interfaceid": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.288691] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 897.288998] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7afd8efa-53c9-402e-8d43-5d01005446ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.296277] env[61768]: DEBUG oslo_vmware.api [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 897.296277] env[61768]: value = "task-1229153" [ 897.296277] env[61768]: _type = "Task" [ 897.296277] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.306394] env[61768]: DEBUG oslo_vmware.api [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229153, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.366843] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.496751] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229150, 'name': ReconfigVM_Task, 'duration_secs': 0.173481} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.497082] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 897.497947] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887d4a14-5a56-49ca-983a-2bf10f23cc6f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.520471] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] e2b7cb22-475b-4b1b-b30e-6061e888691e/e2b7cb22-475b-4b1b-b30e-6061e888691e.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.520789] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-231471d3-4d54-4426-b7ac-e4d0460008c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.539651] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 897.539651] env[61768]: value = "task-1229154" [ 897.539651] env[61768]: _type = "Task" [ 897.539651] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.547648] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229154, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.580623] env[61768]: INFO nova.compute.manager [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] instance snapshotting [ 897.583529] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 897.584303] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8739abee-0f49-4e1b-8b6f-3a02396e6fef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.587293] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bf85a6f8-a7f2-495b-83be-c03a8d7c2636 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.607816] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25d14755-1ad3-4985-8e94-ab7ecef34477 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.610782] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 897.610782] env[61768]: value = "task-1229155" [ 897.610782] env[61768]: _type = "Task" [ 897.610782] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.622413] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229155, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.650996] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.322s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.651562] env[61768]: DEBUG nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.749883] env[61768]: DEBUG oslo_vmware.api [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229152, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.155467} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.750205] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 897.750416] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 897.750627] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 897.750853] env[61768]: INFO nova.compute.manager [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Took 1.10 seconds to destroy the instance on the hypervisor. [ 897.751170] env[61768]: DEBUG oslo.service.loopingcall [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.751422] env[61768]: DEBUG nova.compute.manager [-] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 897.751574] env[61768]: DEBUG nova.network.neutron [-] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 897.763655] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.763985] env[61768]: DEBUG nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Instance network_info: |[{"id": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "address": "fa:16:3e:52:a7:22", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7660d6d5-91", "ovs_interfaceid": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.764431] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:a7:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13e71dbb-4279-427c-b39d-ba5df9895e58', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7660d6d5-91ae-41e4-ba9f-1c09038a38c1', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.772330] env[61768]: DEBUG oslo.service.loopingcall [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.772557] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 897.772785] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-928ea160-ef7b-444a-a8c9-c0ff08487a11 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.794696] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.794696] env[61768]: value = "task-1229156" [ 897.794696] env[61768]: _type = "Task" [ 897.794696] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.804866] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229156, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.807826] env[61768]: DEBUG oslo_vmware.api [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229153, 'name': PowerOffVM_Task, 'duration_secs': 0.175723} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.808154] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 897.808269] env[61768]: DEBUG nova.compute.manager [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.808999] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38c4578-dcab-40d5-81b7-f93c669a0822 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.867618] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.050656] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229154, 'name': ReconfigVM_Task, 'duration_secs': 0.396023} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.050974] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Reconfigured VM instance instance-00000056 to attach disk [datastore2] e2b7cb22-475b-4b1b-b30e-6061e888691e/e2b7cb22-475b-4b1b-b30e-6061e888691e.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.051273] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance 'e2b7cb22-475b-4b1b-b30e-6061e888691e' progress to 50 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 898.113652] env[61768]: DEBUG nova.compute.manager [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Received event network-changed-7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.113905] env[61768]: DEBUG nova.compute.manager [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Refreshing instance network info cache due to event network-changed-7660d6d5-91ae-41e4-ba9f-1c09038a38c1. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.114207] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] Acquiring lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.114373] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] Acquired lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.114601] env[61768]: DEBUG nova.network.neutron [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Refreshing network info cache for port 7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 898.120899] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 898.121260] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-5742946f-149d-4277-b6f1-fa5cdce84954 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.131033] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229155, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.131260] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 898.131260] env[61768]: value = "task-1229158" [ 898.131260] env[61768]: _type = "Task" [ 898.131260] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.140425] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229158, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.156497] env[61768]: DEBUG nova.compute.utils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.158120] env[61768]: DEBUG nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 898.158321] env[61768]: DEBUG nova.network.neutron [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 898.222289] env[61768]: DEBUG nova.policy [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fd5852bc16429aa2904b4351c8249d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7857b5217cc499ca8c9c3b8a5471db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.304720] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229156, 'name': CreateVM_Task, 'duration_secs': 0.442223} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.304916] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 898.305600] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.305777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.306124] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.306378] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52624c59-e4b3-434c-ba7b-f82a710c4d7f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.311033] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 898.311033] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52104427-1d77-1b11-63f7-c1a2be51dd1c" [ 898.311033] env[61768]: _type = "Task" [ 898.311033] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.321189] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52104427-1d77-1b11-63f7-c1a2be51dd1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.321815] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0bd8afb7-de5a-46f6-91c9-7b64a6f2f495 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.547s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.370286] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.477798] env[61768]: DEBUG nova.network.neutron [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Successfully created port: bebae6b0-1f32-43c0-8d08-84e5239ed501 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.527209] env[61768]: DEBUG nova.network.neutron [-] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 898.558397] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b92d50a-9873-4b18-8ec7-37b02e708818 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.579773] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3995509e-3e4f-4ab1-a156-03938076d486 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.597333] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance 'e2b7cb22-475b-4b1b-b30e-6061e888691e' progress to 67 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 898.623436] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229155, 'name': CreateSnapshot_Task, 'duration_secs': 0.775709} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.623715] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 898.624808] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f2166df-1900-4258-9d49-c6059defeb62 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.643344] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229158, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.661665] env[61768]: DEBUG nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.821710] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52104427-1d77-1b11-63f7-c1a2be51dd1c, 'name': SearchDatastore_Task, 'duration_secs': 0.011257} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.822114] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.822311] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.822554] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.822711] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.822898] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.823175] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b4b8d52c-a7ea-4c9a-9819-c0db45e2f97d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.832054] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.832248] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 898.832948] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fabd5146-0012-487a-abc9-b31bd3d38016 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.838152] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 898.838152] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5282146b-182b-98e3-dec3-7e99d24d245d" [ 898.838152] env[61768]: _type = "Task" [ 898.838152] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.845645] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5282146b-182b-98e3-dec3-7e99d24d245d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.868482] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.033978] env[61768]: INFO nova.compute.manager [-] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Took 1.28 seconds to deallocate network for instance. [ 899.146156] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 899.150135] env[61768]: DEBUG nova.network.neutron [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Updated VIF entry in instance network info cache for port 7660d6d5-91ae-41e4-ba9f-1c09038a38c1. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 899.150473] env[61768]: DEBUG nova.network.neutron [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Updating instance_info_cache with network_info: [{"id": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "address": "fa:16:3e:52:a7:22", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7660d6d5-91", "ovs_interfaceid": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.151700] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a64bfbd5-9934-4f89-8d48-a811d38531fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.154512] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229158, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.161431] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 899.161431] env[61768]: value = "task-1229159" [ 899.161431] env[61768]: _type = "Task" [ 899.161431] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.167185] env[61768]: DEBUG nova.network.neutron [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Port 5b4c6981-f46a-4e25-9f9a-51106a3988ac binding to destination host cpu-1 is already ACTIVE {{(pid=61768) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 899.178270] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229159, 'name': CloneVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.285845] env[61768]: DEBUG nova.objects.instance [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.352485] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5282146b-182b-98e3-dec3-7e99d24d245d, 'name': SearchDatastore_Task, 'duration_secs': 0.029154} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.353693] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3766025b-2bca-4284-9170-5b128735f61c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.361305] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 899.361305] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5255cfd4-6021-8786-3f91-b5d451f4ed84" [ 899.361305] env[61768]: _type = "Task" [ 899.361305] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.374472] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5255cfd4-6021-8786-3f91-b5d451f4ed84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.377639] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.544508] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.544791] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.545031] env[61768]: DEBUG nova.objects.instance [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lazy-loading 'resources' on Instance uuid 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.644549] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229158, 'name': CreateSnapshot_Task, 'duration_secs': 1.209759} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.646026] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 899.646026] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f237742-5dda-4c1d-920d-d85750f40678 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.656462] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] Releasing lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.656751] env[61768]: DEBUG nova.compute.manager [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Received event network-vif-deleted-a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.656892] env[61768]: INFO nova.compute.manager [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Neutron deleted interface a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0; detaching it from the instance and deleting it from the info cache [ 899.657088] env[61768]: DEBUG nova.network.neutron [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.672566] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229159, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.677186] env[61768]: DEBUG nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.703650] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.703981] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.704228] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.704601] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.704895] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.705089] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.705316] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.705941] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.705941] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.705941] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.706221] env[61768]: DEBUG nova.virt.hardware [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.706874] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4da5892-de68-4dc0-a353-330473f66975 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.716627] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c7117a-91e7-42d4-8df0-2b30888e7d67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.791079] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 899.791287] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.791530] env[61768]: DEBUG nova.network.neutron [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 899.791765] env[61768]: DEBUG nova.objects.instance [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'info_cache' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.874703] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.881478] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5255cfd4-6021-8786-3f91-b5d451f4ed84, 'name': SearchDatastore_Task, 'duration_secs': 0.014152} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.881762] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.882077] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] d54c94a6-0a53-47a8-b024-1a05439da837/d54c94a6-0a53-47a8-b024-1a05439da837.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.882342] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f838de17-33bd-487e-864c-b1d17ca4d5d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.892381] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 899.892381] env[61768]: value = "task-1229160" [ 899.892381] env[61768]: _type = "Task" [ 899.892381] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.901198] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229160, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.168801] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 900.169912] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a9319ef1-faef-4734-9bca-48905dc20554 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.173995] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94ef3bd5-74c6-48da-9878-daa9cfc6a2ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.193384] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.193679] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.193872] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.211381] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229159, 'name': CloneVM_Task} progress is 95%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.211721] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 900.211721] env[61768]: value = "task-1229162" [ 900.211721] env[61768]: _type = "Task" [ 900.211721] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.215139] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4773e3fe-7f72-49f9-a245-ffd79756c550 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.242846] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229162, 'name': CloneVM_Task} progress is 11%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.265061] env[61768]: DEBUG nova.compute.manager [req-b3e4ce1c-9b33-4010-9801-3e2db64e3dc6 req-44dfab45-09ef-4927-b563-95fcaa2e4b80 service nova] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Detach interface failed, port_id=a5a53f41-9ee5-492e-8cd7-f5a17ee01dd0, reason: Instance 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 900.270806] env[61768]: DEBUG nova.compute.manager [req-baabfe7d-acdf-464c-a00e-4b1b7bafa29f req-45030237-16f1-4930-8944-c175619047af service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Received event network-vif-plugged-bebae6b0-1f32-43c0-8d08-84e5239ed501 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.271096] env[61768]: DEBUG oslo_concurrency.lockutils [req-baabfe7d-acdf-464c-a00e-4b1b7bafa29f req-45030237-16f1-4930-8944-c175619047af service nova] Acquiring lock "739faae0-5811-4f88-b56b-1350c0f7b8be-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.271339] env[61768]: DEBUG oslo_concurrency.lockutils [req-baabfe7d-acdf-464c-a00e-4b1b7bafa29f req-45030237-16f1-4930-8944-c175619047af service nova] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.271519] env[61768]: DEBUG oslo_concurrency.lockutils [req-baabfe7d-acdf-464c-a00e-4b1b7bafa29f req-45030237-16f1-4930-8944-c175619047af service nova] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.271705] env[61768]: DEBUG nova.compute.manager [req-baabfe7d-acdf-464c-a00e-4b1b7bafa29f req-45030237-16f1-4930-8944-c175619047af service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] No waiting events found dispatching network-vif-plugged-bebae6b0-1f32-43c0-8d08-84e5239ed501 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.271878] env[61768]: WARNING nova.compute.manager [req-baabfe7d-acdf-464c-a00e-4b1b7bafa29f req-45030237-16f1-4930-8944-c175619047af service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Received unexpected event network-vif-plugged-bebae6b0-1f32-43c0-8d08-84e5239ed501 for instance with vm_state building and task_state spawning. [ 900.295625] env[61768]: DEBUG nova.objects.base [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Object Instance<7697ecda-ef30-447d-a92a-3f5cb3cc9118> lazy-loaded attributes: flavor,info_cache {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 900.371643] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b229d10e-e398-4f70-8298-9b36dc483676 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.385309] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.389065] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bf6985-a38d-49ea-acb8-d73000e5fe71 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.405207] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229160, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.435650] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0681540e-59dc-4f94-98bb-f7920ddf64b8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.440333] env[61768]: DEBUG nova.network.neutron [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Successfully updated port: bebae6b0-1f32-43c0-8d08-84e5239ed501 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.446031] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2988c5-ec1b-44af-bd88-d07c6668d063 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.465881] env[61768]: DEBUG nova.compute.provider_tree [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.680576] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229159, 'name': CloneVM_Task, 'duration_secs': 1.361494} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.680884] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Created linked-clone VM from snapshot [ 900.681681] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50a50143-0985-49a9-9de6-379c74cea958 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.689619] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Uploading image 31037c49-6340-4bb8-8d74-1b462f8d7c43 {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 900.714170] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 900.714170] env[61768]: value = "vm-265501" [ 900.714170] env[61768]: _type = "VirtualMachine" [ 900.714170] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 900.714483] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-dd15cf2c-e2ad-493b-99a7-7471f5407071 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.725680] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229162, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.727169] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lease: (returnval){ [ 900.727169] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1d4a7-77d5-e221-a4ea-150380da0e47" [ 900.727169] env[61768]: _type = "HttpNfcLease" [ 900.727169] env[61768]: } obtained for exporting VM: (result){ [ 900.727169] env[61768]: value = "vm-265501" [ 900.727169] env[61768]: _type = "VirtualMachine" [ 900.727169] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 900.727503] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the lease: (returnval){ [ 900.727503] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1d4a7-77d5-e221-a4ea-150380da0e47" [ 900.727503] env[61768]: _type = "HttpNfcLease" [ 900.727503] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 900.734107] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 900.734107] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1d4a7-77d5-e221-a4ea-150380da0e47" [ 900.734107] env[61768]: _type = "HttpNfcLease" [ 900.734107] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 900.873728] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.904182] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229160, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.553484} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.904520] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] d54c94a6-0a53-47a8-b024-1a05439da837/d54c94a6-0a53-47a8-b024-1a05439da837.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.904651] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.904881] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e77ea08-7222-4090-a16a-81479d04d628 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.911849] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 900.911849] env[61768]: value = "task-1229164" [ 900.911849] env[61768]: _type = "Task" [ 900.911849] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.922593] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229164, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.943418] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "refresh_cache-739faae0-5811-4f88-b56b-1350c0f7b8be" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.943586] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "refresh_cache-739faae0-5811-4f88-b56b-1350c0f7b8be" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.943840] env[61768]: DEBUG nova.network.neutron [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 900.969692] env[61768]: DEBUG nova.scheduler.client.report [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.021049] env[61768]: DEBUG nova.network.neutron [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating instance_info_cache with network_info: [{"id": "51b2319f-ceb1-41db-8019-1d918e70080d", "address": "fa:16:3e:7d:47:ab", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.157", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51b2319f-ce", "ovs_interfaceid": "51b2319f-ceb1-41db-8019-1d918e70080d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.226563] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229162, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.235460] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 901.235460] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1d4a7-77d5-e221-a4ea-150380da0e47" [ 901.235460] env[61768]: _type = "HttpNfcLease" [ 901.235460] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 901.235774] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 901.235774] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e1d4a7-77d5-e221-a4ea-150380da0e47" [ 901.235774] env[61768]: _type = "HttpNfcLease" [ 901.235774] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 901.237035] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cbdfd95-4120-46bb-b383-d8a574bca782 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.245675] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3211d-1893-6309-ae1e-14d14f7ef677/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 901.245901] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3211d-1893-6309-ae1e-14d14f7ef677/disk-0.vmdk for reading. {{(pid=61768) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 901.313758] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.313993] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.314401] env[61768]: DEBUG nova.network.neutron [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.351141] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-66b6682f-74ee-4b25-b73a-7cca992fb912 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.373972] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.421487] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229164, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082689} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.421782] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.422615] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a087bf8e-1b3f-4509-b4b3-f33831f16e35 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.444683] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] d54c94a6-0a53-47a8-b024-1a05439da837/d54c94a6-0a53-47a8-b024-1a05439da837.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.444942] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d946e2c-e2c7-4f8e-acf2-5b13c1454b67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.466982] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 901.466982] env[61768]: value = "task-1229165" [ 901.466982] env[61768]: _type = "Task" [ 901.466982] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.477940] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.931s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.477940] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.488388] env[61768]: DEBUG nova.network.neutron [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 901.496230] env[61768]: INFO nova.scheduler.client.report [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted allocations for instance 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f [ 901.523606] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "refresh_cache-7697ecda-ef30-447d-a92a-3f5cb3cc9118" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.722098] env[61768]: DEBUG nova.network.neutron [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Updating instance_info_cache with network_info: [{"id": "bebae6b0-1f32-43c0-8d08-84e5239ed501", "address": "fa:16:3e:6d:e4:81", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbebae6b0-1f", "ovs_interfaceid": "bebae6b0-1f32-43c0-8d08-84e5239ed501", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.727458] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229162, 'name': CloneVM_Task} progress is 95%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.874534] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.980720] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229165, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.015757] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3550a84d-c7d2-4a67-9d75-5cb25ee2980a tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "659e8d4c-1af6-4e81-b6a1-2d9466af6b0f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.366s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.027833] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 902.029105] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f48c049-1189-4411-88d9-68132a555a6a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.039529] env[61768]: DEBUG oslo_vmware.api [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 902.039529] env[61768]: value = "task-1229166" [ 902.039529] env[61768]: _type = "Task" [ 902.039529] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.055913] env[61768]: DEBUG oslo_vmware.api [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229166, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.078046] env[61768]: DEBUG nova.network.neutron [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance_info_cache with network_info: [{"id": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "address": "fa:16:3e:2b:a2:c7", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4c6981-f4", "ovs_interfaceid": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.227272] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229162, 'name': CloneVM_Task, 'duration_secs': 1.662381} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.227762] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Created linked-clone VM from snapshot [ 902.228651] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-361901cf-23c4-4ea9-a6dd-304665ba4285 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.231842] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "refresh_cache-739faae0-5811-4f88-b56b-1350c0f7b8be" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.232428] env[61768]: DEBUG nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Instance network_info: |[{"id": "bebae6b0-1f32-43c0-8d08-84e5239ed501", "address": "fa:16:3e:6d:e4:81", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbebae6b0-1f", "ovs_interfaceid": "bebae6b0-1f32-43c0-8d08-84e5239ed501", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 902.232869] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6d:e4:81', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bebae6b0-1f32-43c0-8d08-84e5239ed501', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 902.240713] env[61768]: DEBUG oslo.service.loopingcall [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.241503] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 902.241786] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e6a75d7-bb65-41fc-a052-06f01d9b06c3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.262558] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Uploading image 3e195857-f317-4c61-a558-ef0357b6e879 {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 902.272475] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 902.272475] env[61768]: value = "task-1229167" [ 902.272475] env[61768]: _type = "Task" [ 902.272475] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.282469] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229167, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.283668] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 902.283941] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-750a4ce6-26b4-4ad9-9ef6-11b3378f06b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.288636] env[61768]: DEBUG nova.compute.manager [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Received event network-changed-bebae6b0-1f32-43c0-8d08-84e5239ed501 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.288858] env[61768]: DEBUG nova.compute.manager [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Refreshing instance network info cache due to event network-changed-bebae6b0-1f32-43c0-8d08-84e5239ed501. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.289101] env[61768]: DEBUG oslo_concurrency.lockutils [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] Acquiring lock "refresh_cache-739faae0-5811-4f88-b56b-1350c0f7b8be" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.289261] env[61768]: DEBUG oslo_concurrency.lockutils [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] Acquired lock "refresh_cache-739faae0-5811-4f88-b56b-1350c0f7b8be" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.289428] env[61768]: DEBUG nova.network.neutron [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Refreshing network info cache for port bebae6b0-1f32-43c0-8d08-84e5239ed501 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 902.297272] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 902.297272] env[61768]: value = "task-1229168" [ 902.297272] env[61768]: _type = "Task" [ 902.297272] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.308442] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229168, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.380262] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.413794] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.413794] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.414034] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.414139] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.414358] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.417237] env[61768]: INFO nova.compute.manager [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Terminating instance [ 902.419703] env[61768]: DEBUG nova.compute.manager [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.419963] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 902.420838] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e383ea8-94fe-4bb6-9240-d844c6e2cc9f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.430300] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 902.430713] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6caa10a-fbab-45d1-a3c4-2540745ee404 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.438988] env[61768]: DEBUG oslo_vmware.api [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 902.438988] env[61768]: value = "task-1229169" [ 902.438988] env[61768]: _type = "Task" [ 902.438988] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.449021] env[61768]: DEBUG oslo_vmware.api [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229169, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.479863] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229165, 'name': ReconfigVM_Task, 'duration_secs': 0.610464} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.482120] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Reconfigured VM instance instance-00000059 to attach disk [datastore1] d54c94a6-0a53-47a8-b024-1a05439da837/d54c94a6-0a53-47a8-b024-1a05439da837.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.483228] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-96172647-0468-4ad7-81f5-4f014264b85e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.493855] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 902.493855] env[61768]: value = "task-1229170" [ 902.493855] env[61768]: _type = "Task" [ 902.493855] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.505439] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229170, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.551777] env[61768]: DEBUG oslo_vmware.api [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229166, 'name': PowerOnVM_Task, 'duration_secs': 0.511398} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.552648] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 902.553367] env[61768]: DEBUG nova.compute.manager [None req-7cdee97e-5dc4-435a-b77b-189884df9d76 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.554027] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc570d47-187a-41b3-824c-abcab750c331 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.583055] env[61768]: DEBUG oslo_concurrency.lockutils [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.783880] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229167, 'name': CreateVM_Task, 'duration_secs': 0.453633} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.784072] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 902.784788] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.784980] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.785371] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.785653] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d61e60e6-f097-43e7-adf3-43a499928de6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.795432] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 902.795432] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223ac04-fe08-4a09-fca0-a7aa35f11ace" [ 902.795432] env[61768]: _type = "Task" [ 902.795432] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.808282] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223ac04-fe08-4a09-fca0-a7aa35f11ace, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.811768] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229168, 'name': Destroy_Task, 'duration_secs': 0.496383} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.812047] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Destroyed the VM [ 902.812300] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 902.812596] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4e5ab99c-fdbe-478e-b77e-81e3dc64e1fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.820358] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 902.820358] env[61768]: value = "task-1229171" [ 902.820358] env[61768]: _type = "Task" [ 902.820358] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.830303] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229171, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.878608] env[61768]: DEBUG oslo_vmware.api [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229149, 'name': ReconfigVM_Task, 'duration_secs': 5.810787} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.879096] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.879418] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Reconfigured VM to detach interface {{(pid=61768) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 902.950934] env[61768]: DEBUG oslo_vmware.api [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229169, 'name': PowerOffVM_Task, 'duration_secs': 0.312746} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.951250] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 902.951644] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.951943] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d11cb88d-6a7b-45e8-8863-90ff1dba51cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.006925] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229170, 'name': Rename_Task, 'duration_secs': 0.24037} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.007348] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 903.007528] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b1ec2b6-660d-4fd6-96a8-4f2bc6390709 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.015426] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 903.015426] env[61768]: value = "task-1229173" [ 903.015426] env[61768]: _type = "Task" [ 903.015426] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.027976] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229173, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.044252] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 903.044561] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 903.044787] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleting the datastore file [datastore2] 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.045121] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f36ce9e-7dd7-4e29-8aa5-d35815c747c3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.051500] env[61768]: DEBUG nova.network.neutron [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Updated VIF entry in instance network info cache for port bebae6b0-1f32-43c0-8d08-84e5239ed501. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 903.051909] env[61768]: DEBUG nova.network.neutron [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Updating instance_info_cache with network_info: [{"id": "bebae6b0-1f32-43c0-8d08-84e5239ed501", "address": "fa:16:3e:6d:e4:81", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbebae6b0-1f", "ovs_interfaceid": "bebae6b0-1f32-43c0-8d08-84e5239ed501", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.058447] env[61768]: DEBUG oslo_vmware.api [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for the task: (returnval){ [ 903.058447] env[61768]: value = "task-1229174" [ 903.058447] env[61768]: _type = "Task" [ 903.058447] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.079380] env[61768]: DEBUG oslo_vmware.api [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229174, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.115214] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ccc24a5-ebdf-4390-b1d4-f5e2be4acb7c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.139023] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0cac48-dc71-4bbc-823b-f10e0b6225f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.148236] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance 'e2b7cb22-475b-4b1b-b30e-6061e888691e' progress to 83 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 903.311691] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5223ac04-fe08-4a09-fca0-a7aa35f11ace, 'name': SearchDatastore_Task, 'duration_secs': 0.015453} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.312436] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.313045] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.313624] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.314305] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.314488] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.314915] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9670c7ae-c2a3-44aa-b16a-d53711f27860 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.329032] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.329032] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 903.329032] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1047d36d-db83-419a-99a3-b185b58eea2f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.334390] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229171, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.339089] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 903.339089] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5297b441-1a86-833f-cfcc-2c1fc35e99cb" [ 903.339089] env[61768]: _type = "Task" [ 903.339089] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.351398] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5297b441-1a86-833f-cfcc-2c1fc35e99cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.527120] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229173, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.558109] env[61768]: DEBUG oslo_concurrency.lockutils [req-6a6bd92a-edf9-4da6-8440-86aecd9319e1 req-eb791ca5-c155-4328-bc4a-f2cf9301ec2e service nova] Releasing lock "refresh_cache-739faae0-5811-4f88-b56b-1350c0f7b8be" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.570036] env[61768]: DEBUG oslo_vmware.api [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Task: {'id': task-1229174, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.273706} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.570036] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.570263] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 903.570418] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 903.570621] env[61768]: INFO nova.compute.manager [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Took 1.15 seconds to destroy the instance on the hypervisor. [ 903.570926] env[61768]: DEBUG oslo.service.loopingcall [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.571198] env[61768]: DEBUG nova.compute.manager [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.571293] env[61768]: DEBUG nova.network.neutron [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 903.655915] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 903.655915] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d18e536-2175-46cd-8ba5-26723fd176dc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.666052] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 903.666052] env[61768]: value = "task-1229175" [ 903.666052] env[61768]: _type = "Task" [ 903.666052] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.675470] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229175, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.833408] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229171, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.853924] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5297b441-1a86-833f-cfcc-2c1fc35e99cb, 'name': SearchDatastore_Task, 'duration_secs': 0.017416} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.858138] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0312a75e-5f7a-4ae4-b3fc-364284376473 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.865153] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 903.865153] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524d732a-d01b-fe8e-491f-ca9a2c0cde77" [ 903.865153] env[61768]: _type = "Task" [ 903.865153] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.877534] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524d732a-d01b-fe8e-491f-ca9a2c0cde77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.960427] env[61768]: DEBUG nova.compute.manager [req-ce41b7fb-7e1a-46e0-8b7a-1ae1460c42cc req-9a406d14-17bb-47ef-84b6-bb06f18e13d3 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Received event network-vif-deleted-e2db4501-9ba5-4717-8ce1-2565bb20c603 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.960605] env[61768]: INFO nova.compute.manager [req-ce41b7fb-7e1a-46e0-8b7a-1ae1460c42cc req-9a406d14-17bb-47ef-84b6-bb06f18e13d3 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Neutron deleted interface e2db4501-9ba5-4717-8ce1-2565bb20c603; detaching it from the instance and deleting it from the info cache [ 903.960903] env[61768]: DEBUG nova.network.neutron [req-ce41b7fb-7e1a-46e0-8b7a-1ae1460c42cc req-9a406d14-17bb-47ef-84b6-bb06f18e13d3 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.026435] env[61768]: DEBUG oslo_vmware.api [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229173, 'name': PowerOnVM_Task, 'duration_secs': 0.706238} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.026938] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 904.026938] env[61768]: INFO nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Took 8.57 seconds to spawn the instance on the hypervisor. [ 904.027253] env[61768]: DEBUG nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.028229] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20c72d1-7ce4-4f28-a76e-61409b51360d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.179703] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229175, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.324322] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Acquiring lock "30f55051-63b3-47a9-83fb-945cdeb82574" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.324556] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "30f55051-63b3-47a9-83fb-945cdeb82574" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.335848] env[61768]: DEBUG oslo_vmware.api [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229171, 'name': RemoveSnapshot_Task, 'duration_secs': 1.015704} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.336649] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 904.358945] env[61768]: DEBUG nova.network.neutron [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.377039] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524d732a-d01b-fe8e-491f-ca9a2c0cde77, 'name': SearchDatastore_Task, 'duration_secs': 0.017513} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.377578] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.377842] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 739faae0-5811-4f88-b56b-1350c0f7b8be/739faae0-5811-4f88-b56b-1350c0f7b8be.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 904.378115] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a73a55ee-5e34-4401-9175-85a3d5f62b38 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.386400] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 904.386400] env[61768]: value = "task-1229176" [ 904.386400] env[61768]: _type = "Task" [ 904.386400] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.395825] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229176, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.416027] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.416249] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.416436] env[61768]: DEBUG nova.network.neutron [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 904.464007] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-78dce607-3e3a-4b67-a4f6-56ebe04de321 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.474942] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11aecd89-afad-4bcd-bd82-234471d1fcf7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.513844] env[61768]: DEBUG nova.compute.manager [req-ce41b7fb-7e1a-46e0-8b7a-1ae1460c42cc req-9a406d14-17bb-47ef-84b6-bb06f18e13d3 service nova] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Detach interface failed, port_id=e2db4501-9ba5-4717-8ce1-2565bb20c603, reason: Instance 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 904.548353] env[61768]: INFO nova.compute.manager [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Took 16.22 seconds to build instance. [ 904.680747] env[61768]: DEBUG oslo_vmware.api [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229175, 'name': PowerOnVM_Task, 'duration_secs': 0.514937} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.681119] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 904.681325] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-90a3d520-3a8b-4964-b1ec-d66cb91bfefe tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance 'e2b7cb22-475b-4b1b-b30e-6061e888691e' progress to 100 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 904.830112] env[61768]: DEBUG nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 904.841541] env[61768]: WARNING nova.compute.manager [None req-40996661-9269-49e4-b0d0-9b45e7754f70 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Image not found during snapshot: nova.exception.ImageNotFound: Image 3e195857-f317-4c61-a558-ef0357b6e879 could not be found. [ 904.861870] env[61768]: INFO nova.compute.manager [-] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Took 1.29 seconds to deallocate network for instance. [ 904.900022] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229176, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.966115] env[61768]: DEBUG nova.compute.manager [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Received event network-changed-7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.966115] env[61768]: DEBUG nova.compute.manager [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Refreshing instance network info cache due to event network-changed-7660d6d5-91ae-41e4-ba9f-1c09038a38c1. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 904.966239] env[61768]: DEBUG oslo_concurrency.lockutils [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] Acquiring lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.966445] env[61768]: DEBUG oslo_concurrency.lockutils [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] Acquired lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.966631] env[61768]: DEBUG nova.network.neutron [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Refreshing network info cache for port 7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 905.052159] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2f358459-4cdb-47f4-a3ac-aab31f7030d8 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "d54c94a6-0a53-47a8-b024-1a05439da837" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.732s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.317947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.317947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.317947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.317947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.317947] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.319078] env[61768]: INFO nova.compute.manager [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Terminating instance [ 905.321726] env[61768]: DEBUG nova.compute.manager [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.322139] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 905.323149] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a20d646-fc4d-4d3c-af56-610a158efb11 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.334088] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.334088] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8aae96bd-5371-495c-88d5-c94b099e1eb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.343079] env[61768]: DEBUG oslo_vmware.api [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 905.343079] env[61768]: value = "task-1229177" [ 905.343079] env[61768]: _type = "Task" [ 905.343079] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.357527] env[61768]: DEBUG oslo_vmware.api [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229177, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.370125] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.370125] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.371653] env[61768]: INFO nova.compute.claims [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.380028] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.396598] env[61768]: INFO nova.network.neutron [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Port 20999117-9aaf-4291-9e72-f8304dfdf3f0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 905.398361] env[61768]: DEBUG nova.network.neutron [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.404933] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229176, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.605465} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.405809] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 739faae0-5811-4f88-b56b-1350c0f7b8be/739faae0-5811-4f88-b56b-1350c0f7b8be.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 905.405809] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 905.405809] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04bef2d9-a586-49bc-8535-28233bcc1b32 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.413828] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 905.413828] env[61768]: value = "task-1229178" [ 905.413828] env[61768]: _type = "Task" [ 905.413828] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.426571] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.697469] env[61768]: DEBUG nova.network.neutron [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Updated VIF entry in instance network info cache for port 7660d6d5-91ae-41e4-ba9f-1c09038a38c1. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.697469] env[61768]: DEBUG nova.network.neutron [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Updating instance_info_cache with network_info: [{"id": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "address": "fa:16:3e:52:a7:22", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7660d6d5-91", "ovs_interfaceid": "7660d6d5-91ae-41e4-ba9f-1c09038a38c1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.852428] env[61768]: DEBUG oslo_vmware.api [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229177, 'name': PowerOffVM_Task, 'duration_secs': 0.330473} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.852761] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 905.852946] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 905.853226] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-228aa9c0-4a43-4e43-966a-a1accfd82755 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.901699] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.924196] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157242} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.924482] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.925312] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a9164e-43ea-433a-a3c0-0dc35189ba1e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.953553] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 739faae0-5811-4f88-b56b-1350c0f7b8be/739faae0-5811-4f88-b56b-1350c0f7b8be.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.956235] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2dc8228c-43ab-4ce1-8e0c-8119daa76719 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.971981] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 905.972342] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 905.972581] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleting the datastore file [datastore2] aaf205c5-dd11-4d1c-90bd-3ecd5a121227 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.972976] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-131c67e4-0295-475b-ab64-2577a3203dd5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.981398] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 905.981398] env[61768]: value = "task-1229181" [ 905.981398] env[61768]: _type = "Task" [ 905.981398] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.982942] env[61768]: DEBUG oslo_vmware.api [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for the task: (returnval){ [ 905.982942] env[61768]: value = "task-1229180" [ 905.982942] env[61768]: _type = "Task" [ 905.982942] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.993886] env[61768]: DEBUG nova.compute.manager [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.994416] env[61768]: DEBUG nova.compute.manager [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing instance network info cache due to event network-changed-b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.994416] env[61768]: DEBUG oslo_concurrency.lockutils [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] Acquiring lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.994576] env[61768]: DEBUG oslo_concurrency.lockutils [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] Acquired lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.994798] env[61768]: DEBUG nova.network.neutron [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Refreshing network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 906.000553] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229181, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.007479] env[61768]: DEBUG oslo_vmware.api [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.117739] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-5cc705a5-1423-43e1-8356-9abd67a8da3a-20999117-9aaf-4291-9e72-f8304dfdf3f0" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.118227] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-5cc705a5-1423-43e1-8356-9abd67a8da3a-20999117-9aaf-4291-9e72-f8304dfdf3f0" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.118421] env[61768]: DEBUG nova.objects.instance [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'flavor' on Instance uuid 5cc705a5-1423-43e1-8356-9abd67a8da3a {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.205076] env[61768]: DEBUG oslo_concurrency.lockutils [req-4669edeb-6a62-4105-8664-40de5626d342 req-c3f1d44b-089e-4a0b-a7bc-088f392730d3 service nova] Releasing lock "refresh_cache-d54c94a6-0a53-47a8-b024-1a05439da837" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.406609] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c4dbd2c2-afb0-41a5-b6b0-993517ce7ed7 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-ecd08348-b22f-49c5-b13f-b187506e38d6-20999117-9aaf-4291-9e72-f8304dfdf3f0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.124s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.502556] env[61768]: DEBUG oslo_vmware.api [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Task: {'id': task-1229180, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.277937} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.504628] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229181, 'name': ReconfigVM_Task, 'duration_secs': 0.426587} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.504925] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.505162] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 906.505369] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.505567] env[61768]: INFO nova.compute.manager [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Took 1.18 seconds to destroy the instance on the hypervisor. [ 906.505849] env[61768]: DEBUG oslo.service.loopingcall [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.506127] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 739faae0-5811-4f88-b56b-1350c0f7b8be/739faae0-5811-4f88-b56b-1350c0f7b8be.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.509676] env[61768]: DEBUG nova.compute.manager [-] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 906.509676] env[61768]: DEBUG nova.network.neutron [-] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 906.511643] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-31521aed-a02e-4e87-91d9-1240e02cfc36 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.521197] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 906.521197] env[61768]: value = "task-1229182" [ 906.521197] env[61768]: _type = "Task" [ 906.521197] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.533516] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229182, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.639479] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622aa7a6-6793-4399-b353-289e9bcc6733 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.649628] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ecad5e-ba10-4c0b-bb22-20c9e5654a6a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.690068] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2075d3-5ffc-4e4e-99c3-b2f965dbc508 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.700135] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8991dd4-bcd8-4e87-8114-8435943c314d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.720766] env[61768]: DEBUG nova.compute.provider_tree [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.820968] env[61768]: DEBUG nova.objects.instance [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'pci_requests' on Instance uuid 5cc705a5-1423-43e1-8356-9abd67a8da3a {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.834341] env[61768]: DEBUG nova.network.neutron [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updated VIF entry in instance network info cache for port b4ad17b4-d077-472c-be99-bd42736950e4. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.834766] env[61768]: DEBUG nova.network.neutron [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [{"id": "b4ad17b4-d077-472c-be99-bd42736950e4", "address": "fa:16:3e:65:99:15", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4ad17b4-d0", "ovs_interfaceid": "b4ad17b4-d077-472c-be99-bd42736950e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.021506] env[61768]: DEBUG nova.compute.manager [req-db0617a6-5884-4ed7-9104-2b81189b602d req-ab249a8b-76f0-4c16-89ee-fcf74d1128be service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Received event network-vif-deleted-7084aac0-6dd6-411e-84e9-e4a2e353429c {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.021752] env[61768]: INFO nova.compute.manager [req-db0617a6-5884-4ed7-9104-2b81189b602d req-ab249a8b-76f0-4c16-89ee-fcf74d1128be service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Neutron deleted interface 7084aac0-6dd6-411e-84e9-e4a2e353429c; detaching it from the instance and deleting it from the info cache [ 907.021939] env[61768]: DEBUG nova.network.neutron [req-db0617a6-5884-4ed7-9104-2b81189b602d req-ab249a8b-76f0-4c16-89ee-fcf74d1128be service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.034308] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229182, 'name': Rename_Task, 'duration_secs': 0.202088} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.035176] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 907.035459] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fad0b047-59f6-43ef-bbc5-6517f0ed52d0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.044385] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 907.044385] env[61768]: value = "task-1229183" [ 907.044385] env[61768]: _type = "Task" [ 907.044385] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.053877] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229183, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.224890] env[61768]: DEBUG nova.scheduler.client.report [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.323311] env[61768]: DEBUG nova.objects.base [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Object Instance<5cc705a5-1423-43e1-8356-9abd67a8da3a> lazy-loaded attributes: flavor,pci_requests {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 907.323571] env[61768]: DEBUG nova.network.neutron [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 907.337649] env[61768]: DEBUG oslo_concurrency.lockutils [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] Releasing lock "refresh_cache-ecd08348-b22f-49c5-b13f-b187506e38d6" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.338035] env[61768]: DEBUG nova.compute.manager [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.338445] env[61768]: DEBUG nova.compute.manager [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing instance network info cache due to event network-changed-604a3820-78ee-48e4-bb9e-af1273963ff8. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 907.338506] env[61768]: DEBUG oslo_concurrency.lockutils [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.338692] env[61768]: DEBUG oslo_concurrency.lockutils [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.338898] env[61768]: DEBUG nova.network.neutron [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing network info cache for port 604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 907.355559] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.355803] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.356048] env[61768]: DEBUG nova.compute.manager [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Going to confirm migration 2 {{(pid=61768) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 907.377050] env[61768]: DEBUG nova.network.neutron [-] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.399663] env[61768]: DEBUG nova.policy [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc14f7d2d2724726890c1101a784e07f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2f3e8b04445948a6a8a737ad89a5c002', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.524812] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-11ed3198-9519-4d0e-8442-e994beb4c65e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.535957] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf76d085-3956-4f08-b7d3-072c37594ac2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.558798] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229183, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.574810] env[61768]: DEBUG nova.compute.manager [req-db0617a6-5884-4ed7-9104-2b81189b602d req-ab249a8b-76f0-4c16-89ee-fcf74d1128be service nova] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Detach interface failed, port_id=7084aac0-6dd6-411e-84e9-e4a2e353429c, reason: Instance aaf205c5-dd11-4d1c-90bd-3ecd5a121227 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 907.731772] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.732264] env[61768]: DEBUG nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 907.735213] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.359s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.735453] env[61768]: DEBUG nova.objects.instance [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lazy-loading 'resources' on Instance uuid 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.879897] env[61768]: INFO nova.compute.manager [-] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Took 1.37 seconds to deallocate network for instance. [ 907.897881] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.898165] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.898357] env[61768]: DEBUG nova.network.neutron [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 907.898523] env[61768]: DEBUG nova.objects.instance [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lazy-loading 'info_cache' on Instance uuid e2b7cb22-475b-4b1b-b30e-6061e888691e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.060994] env[61768]: DEBUG oslo_vmware.api [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229183, 'name': PowerOnVM_Task, 'duration_secs': 0.891098} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.061286] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 908.061493] env[61768]: INFO nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Took 8.38 seconds to spawn the instance on the hypervisor. [ 908.061675] env[61768]: DEBUG nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.062473] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db635039-d7d1-4918-b552-6d9887c1102d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.162884] env[61768]: DEBUG nova.network.neutron [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updated VIF entry in instance network info cache for port 604a3820-78ee-48e4-bb9e-af1273963ff8. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 908.163292] env[61768]: DEBUG nova.network.neutron [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.238398] env[61768]: DEBUG nova.compute.utils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.242700] env[61768]: DEBUG nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.242876] env[61768]: DEBUG nova.network.neutron [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 908.313490] env[61768]: DEBUG nova.policy [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6488600b4444c1d94d56fbedc89b07a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f958bb192d0349c4ac4f256db207d575', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 908.387726] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.453298] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457e595c-e653-443c-9d61-dc27054213d2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.457142] env[61768]: INFO nova.compute.manager [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Rebuilding instance [ 908.465077] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105e0aa0-4f92-445a-8450-b4c7282c3b5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.504227] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8ecf90-1812-4af9-be8f-38b24c9d0238 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.514496] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d9db18-818f-49b1-ab9b-cfe2206e6438 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.524404] env[61768]: DEBUG nova.compute.manager [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.525247] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb4b00b-d8a7-42f6-96a6-c1336bf719a5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.536143] env[61768]: DEBUG nova.compute.provider_tree [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 908.582586] env[61768]: INFO nova.compute.manager [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Took 17.17 seconds to build instance. [ 908.666098] env[61768]: DEBUG oslo_concurrency.lockutils [req-7cc011d2-695d-435f-a16e-be359b66f7b1 req-c106a913-2845-4b96-b095-0d03af46ffa9 service nova] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.744193] env[61768]: DEBUG nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 908.818874] env[61768]: DEBUG nova.network.neutron [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Successfully created port: 4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 908.947046] env[61768]: DEBUG nova.network.neutron [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Successfully updated port: 20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 909.041018] env[61768]: DEBUG nova.scheduler.client.report [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 909.046229] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.052032] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fccb63ce-dcc4-4436-99f5-6e88ab5e46cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.061311] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 909.061311] env[61768]: value = "task-1229184" [ 909.061311] env[61768]: _type = "Task" [ 909.061311] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.076609] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229184, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.090955] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2df6cc07-a69a-4cd1-8d87-266edab8e9ff tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.694s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.179918] env[61768]: DEBUG nova.network.neutron [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance_info_cache with network_info: [{"id": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "address": "fa:16:3e:2b:a2:c7", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b4c6981-f4", "ovs_interfaceid": "5b4c6981-f46a-4e25-9f9a-51106a3988ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.251074] env[61768]: INFO nova.virt.block_device [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Booting with volume 68c9b861-d548-4ed9-b618-463dfbd8cc16 at /dev/sda [ 909.296135] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-27c9bbf3-acf6-48cd-9d49-f680df8fd815 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.310021] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df72da46-383f-4fc3-98d2-bb23c7c50189 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.349364] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7df52bb7-bcec-49ae-a4ab-0088902b5ef9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.360482] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e430722a-1393-428c-ac72-a0b04c482e92 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.400168] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70f4e2b6-0887-4eb3-8416-e97dd02f6caf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.407954] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65cc6304-887d-4b0f-8c7b-8aad00b3c9e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.431999] env[61768]: DEBUG nova.virt.block_device [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Updating existing volume attachment record: 16676694-049d-496b-9fc4-f84cdd4f26da {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 909.451824] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.451824] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.451824] env[61768]: DEBUG nova.network.neutron [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 909.493469] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "e0c89fa6-9d32-4343-89d7-e9129d252b40" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.493730] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.545870] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.548773] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.161s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.548773] env[61768]: DEBUG nova.objects.instance [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lazy-loading 'resources' on Instance uuid aaf205c5-dd11-4d1c-90bd-3ecd5a121227 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.570368] env[61768]: INFO nova.scheduler.client.report [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Deleted allocations for instance 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54 [ 909.573925] env[61768]: DEBUG nova.compute.manager [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-vif-plugged-20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.574031] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] Acquiring lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.575010] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.575010] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.575010] env[61768]: DEBUG nova.compute.manager [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] No waiting events found dispatching network-vif-plugged-20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 909.575010] env[61768]: WARNING nova.compute.manager [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received unexpected event network-vif-plugged-20999117-9aaf-4291-9e72-f8304dfdf3f0 for instance with vm_state active and task_state None. [ 909.575179] env[61768]: DEBUG nova.compute.manager [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-changed-20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.575219] env[61768]: DEBUG nova.compute.manager [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing instance network info cache due to event network-changed-20999117-9aaf-4291-9e72-f8304dfdf3f0. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.575393] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.584926] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229184, 'name': PowerOffVM_Task, 'duration_secs': 0.27351} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.585081] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 909.585774] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 909.586060] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a327f2e0-3077-4046-a07c-2d73f93fcb70 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.594339] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 909.594339] env[61768]: value = "task-1229185" [ 909.594339] env[61768]: _type = "Task" [ 909.594339] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.603762] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229185, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.681800] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-e2b7cb22-475b-4b1b-b30e-6061e888691e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.682130] env[61768]: DEBUG nova.objects.instance [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lazy-loading 'migration_context' on Instance uuid e2b7cb22-475b-4b1b-b30e-6061e888691e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 909.990509] env[61768]: WARNING nova.network.neutron [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] 2fbe88f9-1bca-4347-914d-a62dc6bebde8 already exists in list: networks containing: ['2fbe88f9-1bca-4347-914d-a62dc6bebde8']. ignoring it [ 909.995561] env[61768]: DEBUG nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 910.082925] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b288eb79-d162-4412-acd6-218be6f3be61 tempest-ServersAdminTestJSON-828081596 tempest-ServersAdminTestJSON-828081596-project-member] Lock "9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.669s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.111156] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 910.111440] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 910.111687] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265469', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'name': 'volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e', 'attached_at': '', 'detached_at': '', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'serial': '9ddc3f0e-f5f4-449a-878f-bda8467085c3'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 910.112509] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fb30dbf-d767-4551-b3bf-df9166242c5e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.144525] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fdf21c-ccbe-4f07-b03b-0783d6affea0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.153887] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883613fc-2bf4-40b7-bc5f-c23a99561ba6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.177079] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-747591a3-e63a-44fe-8740-186e5bcba8c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.185060] env[61768]: DEBUG nova.objects.base [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 910.198539] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97466528-3c64-400b-a440-09af5fa93155 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.202028] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] The volume has not been displaced from its original location: [datastore2] volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3/volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 910.207470] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Reconfiguring VM instance instance-00000053 to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 910.210115] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5d7c456d-33f2-47c3-b77c-2139b94c9e44 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.247155] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2bbd760-71eb-45e6-a42e-05087df7a9b4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.249810] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 910.249810] env[61768]: value = "task-1229186" [ 910.249810] env[61768]: _type = "Task" [ 910.249810] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.258010] env[61768]: DEBUG oslo_vmware.api [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 910.258010] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5211c9fe-d580-bc1a-725d-4f15e0aace95" [ 910.258010] env[61768]: _type = "Task" [ 910.258010] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.265600] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229186, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.275482] env[61768]: DEBUG oslo_vmware.api [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5211c9fe-d580-bc1a-725d-4f15e0aace95, 'name': SearchDatastore_Task, 'duration_secs': 0.011234} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.275838] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.339870] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2470b4ad-5d29-4410-a8fd-c415953a964c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.348585] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399a9c14-00fb-4bb7-a9bd-76f35fd2627f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.379333] env[61768]: DEBUG nova.network.neutron [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "address": "fa:16:3e:ad:82:00", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20999117-9a", "ovs_interfaceid": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.381125] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5011a594-3cfc-4120-b68c-6307cb3705bb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.391798] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f4c5830-5658-446b-8379-4fcc7ceeab48 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.409900] env[61768]: DEBUG nova.compute.provider_tree [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.520846] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.764409] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229186, 'name': ReconfigVM_Task, 'duration_secs': 0.382489} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.764914] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Reconfigured VM instance instance-00000053 to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 910.770854] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03268004-3aba-4c9f-a004-3cb9a5941b32 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.790676] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 910.790676] env[61768]: value = "task-1229187" [ 910.790676] env[61768]: _type = "Task" [ 910.790676] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.801641] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229187, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.804921] env[61768]: DEBUG nova.compute.manager [req-641c0f5f-0fdf-43bb-bd12-6dd8b7bf7a7e req-10302bda-7d12-4746-a84b-d8e71333f68e service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Received event network-vif-plugged-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.805329] env[61768]: DEBUG oslo_concurrency.lockutils [req-641c0f5f-0fdf-43bb-bd12-6dd8b7bf7a7e req-10302bda-7d12-4746-a84b-d8e71333f68e service nova] Acquiring lock "30f55051-63b3-47a9-83fb-945cdeb82574-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.805495] env[61768]: DEBUG oslo_concurrency.lockutils [req-641c0f5f-0fdf-43bb-bd12-6dd8b7bf7a7e req-10302bda-7d12-4746-a84b-d8e71333f68e service nova] Lock "30f55051-63b3-47a9-83fb-945cdeb82574-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.805675] env[61768]: DEBUG oslo_concurrency.lockutils [req-641c0f5f-0fdf-43bb-bd12-6dd8b7bf7a7e req-10302bda-7d12-4746-a84b-d8e71333f68e service nova] Lock "30f55051-63b3-47a9-83fb-945cdeb82574-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.805888] env[61768]: DEBUG nova.compute.manager [req-641c0f5f-0fdf-43bb-bd12-6dd8b7bf7a7e req-10302bda-7d12-4746-a84b-d8e71333f68e service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] No waiting events found dispatching network-vif-plugged-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.806104] env[61768]: WARNING nova.compute.manager [req-641c0f5f-0fdf-43bb-bd12-6dd8b7bf7a7e req-10302bda-7d12-4746-a84b-d8e71333f68e service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Received unexpected event network-vif-plugged-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 for instance with vm_state building and task_state block_device_mapping. [ 910.874816] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3211d-1893-6309-ae1e-14d14f7ef677/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 910.875751] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2e3208-d680-44e0-9506-8bceb7cf679a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.882718] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3211d-1893-6309-ae1e-14d14f7ef677/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 910.882877] env[61768]: ERROR oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3211d-1893-6309-ae1e-14d14f7ef677/disk-0.vmdk due to incomplete transfer. [ 910.883125] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b705bbed-d2b4-4050-93a7-92878263dd99 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.884973] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.885582] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.885746] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.886017] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.886255] env[61768]: DEBUG nova.network.neutron [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Refreshing network info cache for port 20999117-9aaf-4291-9e72-f8304dfdf3f0 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 910.887799] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d547da8e-001d-40a2-8749-52546c9cbbcb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.905580] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.905847] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.906016] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.906261] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.906362] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.906519] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.906732] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.906901] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.907089] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.907266] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.907448] env[61768]: DEBUG nova.virt.hardware [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.913731] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Reconfiguring VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 910.916282] env[61768]: DEBUG nova.scheduler.client.report [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.919208] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16ef7410-d8d1-47c1-90c7-bf1231e5f09c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.931570] env[61768]: DEBUG oslo_vmware.rw_handles [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c3211d-1893-6309-ae1e-14d14f7ef677/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 910.932015] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Uploaded image 31037c49-6340-4bb8-8d74-1b462f8d7c43 to the Glance image server {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 910.934435] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 910.935467] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-647d6c1f-879d-4a66-899b-6e871f466c64 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.944364] env[61768]: DEBUG oslo_vmware.api [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 910.944364] env[61768]: value = "task-1229188" [ 910.944364] env[61768]: _type = "Task" [ 910.944364] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.946192] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 910.946192] env[61768]: value = "task-1229189" [ 910.946192] env[61768]: _type = "Task" [ 910.946192] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.958206] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229189, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.961131] env[61768]: DEBUG oslo_vmware.api [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229188, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.264724] env[61768]: DEBUG nova.network.neutron [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Successfully updated port: 4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.305432] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229187, 'name': ReconfigVM_Task, 'duration_secs': 0.270701} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.305820] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265469', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'name': 'volume-9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e', 'attached_at': '', 'detached_at': '', 'volume_id': '9ddc3f0e-f5f4-449a-878f-bda8467085c3', 'serial': '9ddc3f0e-f5f4-449a-878f-bda8467085c3'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 911.306134] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 911.306998] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b7d5aa5-591e-4027-9604-121ad488905f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.316163] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 911.317127] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bcde07d2-4a7c-4449-b10d-bbc713a6a710 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.319800] env[61768]: DEBUG nova.compute.manager [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Received event network-changed-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.320065] env[61768]: DEBUG nova.compute.manager [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Refreshing instance network info cache due to event network-changed-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 911.320330] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] Acquiring lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.320673] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] Acquired lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.320913] env[61768]: DEBUG nova.network.neutron [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Refreshing network info cache for port 4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 911.391308] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 911.391509] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 911.391722] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Deleting the datastore file [datastore2] c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 911.394508] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-36391640-5c48-4e85-96c5-90753fdf0687 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.406130] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for the task: (returnval){ [ 911.406130] env[61768]: value = "task-1229191" [ 911.406130] env[61768]: _type = "Task" [ 911.406130] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.417549] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229191, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.437662] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.889s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.440211] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.164s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.461786] env[61768]: DEBUG oslo_vmware.api [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229188, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.465174] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229189, 'name': Destroy_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.479448] env[61768]: INFO nova.scheduler.client.report [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Deleted allocations for instance aaf205c5-dd11-4d1c-90bd-3ecd5a121227 [ 911.559318] env[61768]: DEBUG nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 911.559963] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.560221] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.560387] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.560579] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.560759] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.560923] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.561616] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.561616] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.561616] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.561810] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.561850] env[61768]: DEBUG nova.virt.hardware [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.563065] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e058cb-a8d5-499a-af21-fc92ff7f94ee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.571791] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec0523e-6690-4134-9bc0-a139e0c0c998 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.672188] env[61768]: DEBUG nova.network.neutron [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updated VIF entry in instance network info cache for port 20999117-9aaf-4291-9e72-f8304dfdf3f0. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 911.672188] env[61768]: DEBUG nova.network.neutron [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "address": "fa:16:3e:ad:82:00", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20999117-9a", "ovs_interfaceid": "20999117-9aaf-4291-9e72-f8304dfdf3f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 911.768208] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Acquiring lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.866557] env[61768]: DEBUG nova.network.neutron [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 911.919449] env[61768]: DEBUG oslo_vmware.api [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Task: {'id': task-1229191, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.260365} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.919449] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.919449] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 911.919449] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 911.963179] env[61768]: DEBUG oslo_vmware.api [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229188, 'name': ReconfigVM_Task, 'duration_secs': 0.927498} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.969402] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.969744] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Reconfigured VM to attach interface {{(pid=61768) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 911.972373] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229189, 'name': Destroy_Task, 'duration_secs': 0.614778} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.973069] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Destroyed the VM [ 911.973316] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 911.976077] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-0c2f6650-9243-4fc5-a6af-403c7ea4ffc5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.000265] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 912.000265] env[61768]: value = "task-1229192" [ 912.000265] env[61768]: _type = "Task" [ 912.000265] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.001658] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef1dbad0-f824-4081-80ca-338605b07ad5 tempest-ImagesTestJSON-841190651 tempest-ImagesTestJSON-841190651-project-member] Lock "aaf205c5-dd11-4d1c-90bd-3ecd5a121227" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.685s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.014362] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229192, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.019624] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 912.019703] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7ddcf70-1367-4b27-8c5e-98aa4a0da7fc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.032695] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a4f8b9e-7d16-4f9b-a423-54f1f254621f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.049503] env[61768]: DEBUG nova.network.neutron [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.073107] env[61768]: ERROR nova.compute.manager [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Failed to detach volume 9ddc3f0e-f5f4-449a-878f-bda8467085c3 from /dev/sda: nova.exception.InstanceNotFound: Instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e could not be found. [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Traceback (most recent call last): [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self.driver.rebuild(**kwargs) [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] raise NotImplementedError() [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] NotImplementedError [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] During handling of the above exception, another exception occurred: [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Traceback (most recent call last): [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self.driver.detach_volume(context, old_connection_info, [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] return self._volumeops.detach_volume(connection_info, instance) [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self._detach_volume_vmdk(connection_info, instance) [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] stable_ref.fetch_moref(session) [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] raise exception.InstanceNotFound(instance_id=self._uuid) [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] nova.exception.InstanceNotFound: Instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e could not be found. [ 912.073107] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.174098] env[61768]: DEBUG oslo_concurrency.lockutils [req-ed0f83b9-8579-4872-867d-7fdf4e9dbebd req-8c327ab9-b0f8-4565-8e07-9aecb7f1dd40 service nova] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.224156] env[61768]: DEBUG nova.compute.utils [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Build of instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e aborted: Failed to rebuild volume backed instance. {{(pid=61768) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 912.226512] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4f6807d-19ef-4919-a584-f75efd23d2eb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.230715] env[61768]: ERROR nova.compute.manager [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e aborted: Failed to rebuild volume backed instance. [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Traceback (most recent call last): [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self.driver.rebuild(**kwargs) [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] raise NotImplementedError() [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] NotImplementedError [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] During handling of the above exception, another exception occurred: [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Traceback (most recent call last): [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self._detach_root_volume(context, instance, root_bdm) [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] with excutils.save_and_reraise_exception(): [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self.force_reraise() [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] raise self.value [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self.driver.detach_volume(context, old_connection_info, [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] return self._volumeops.detach_volume(connection_info, instance) [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self._detach_volume_vmdk(connection_info, instance) [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] stable_ref.fetch_moref(session) [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] raise exception.InstanceNotFound(instance_id=self._uuid) [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] nova.exception.InstanceNotFound: Instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e could not be found. [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] During handling of the above exception, another exception occurred: [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Traceback (most recent call last): [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] yield [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 912.230715] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self._do_rebuild_instance_with_claim( [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self._do_rebuild_instance( [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self._rebuild_default_impl(**kwargs) [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] self._rebuild_volume_backed_instance( [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] raise exception.BuildAbortException( [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] nova.exception.BuildAbortException: Build of instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e aborted: Failed to rebuild volume backed instance. [ 912.232930] env[61768]: ERROR nova.compute.manager [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] [ 912.240256] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69a02201-8e72-42f7-b987-bfe934001c05 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.283122] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0da71e4b-5ca5-464e-bb2e-3db76532b458 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.294077] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff2d552-7e8c-4779-9d27-24481429f5e3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.308105] env[61768]: DEBUG nova.compute.provider_tree [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.474926] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4c251ebd-4bd5-412d-8c13-1c45cedabe80 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-5cc705a5-1423-43e1-8356-9abd67a8da3a-20999117-9aaf-4291-9e72-f8304dfdf3f0" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.357s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.511223] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229192, 'name': RemoveSnapshot_Task, 'duration_secs': 0.388843} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.511512] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 912.511791] env[61768]: DEBUG nova.compute.manager [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 912.512612] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ba2d09-e79e-44d1-8e0e-e323296fc70c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.554483] env[61768]: DEBUG oslo_concurrency.lockutils [req-cbd69d7e-091a-4fde-8cf0-ada8258dc7c6 req-f294bf2e-4f32-45ff-b132-118bb81464cc service nova] Releasing lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.554839] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Acquired lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.555009] env[61768]: DEBUG nova.network.neutron [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 912.812010] env[61768]: DEBUG nova.scheduler.client.report [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 913.028700] env[61768]: INFO nova.compute.manager [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Shelve offloading [ 913.034849] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 913.034849] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a40619f4-72f7-4480-b469-086ebf5e55ca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.046679] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 913.046679] env[61768]: value = "task-1229193" [ 913.046679] env[61768]: _type = "Task" [ 913.046679] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.059132] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 913.059853] env[61768]: DEBUG nova.compute.manager [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 913.060729] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dddb3f-0df9-429b-a9f8-35817258aa33 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.068718] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.068997] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.069384] env[61768]: DEBUG nova.network.neutron [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.105031] env[61768]: DEBUG nova.network.neutron [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.382839] env[61768]: DEBUG nova.network.neutron [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Updating instance_info_cache with network_info: [{"id": "4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3", "address": "fa:16:3e:f3:d1:42", "network": {"id": "24ffcec7-794b-4e30-984c-4ccb9657d666", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-301275445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f958bb192d0349c4ac4f256db207d575", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd7e358-0e", "ovs_interfaceid": "4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 913.828173] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.388s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.831741] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.311s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.833469] env[61768]: INFO nova.compute.claims [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.887766] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Releasing lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.888129] env[61768]: DEBUG nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Instance network_info: |[{"id": "4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3", "address": "fa:16:3e:f3:d1:42", "network": {"id": "24ffcec7-794b-4e30-984c-4ccb9657d666", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-301275445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f958bb192d0349c4ac4f256db207d575", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd7e358-0e", "ovs_interfaceid": "4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.889229] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:d1:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cd098b1c-636f-492d-b5ae-037cb0cae454', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.897738] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Creating folder: Project (f958bb192d0349c4ac4f256db207d575). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 913.898725] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43c69c8a-08da-49ce-b117-8e6fda63648c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.916458] env[61768]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 913.916637] env[61768]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=61768) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 913.917031] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Folder already exists: Project (f958bb192d0349c4ac4f256db207d575). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.917263] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Creating folder: Instances. Parent ref: group-v265493. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 913.917522] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddcb4fb7-2c7c-43d6-a013-5b2efb435acb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.934184] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Created folder: Instances in parent group-v265493. [ 913.934526] env[61768]: DEBUG oslo.service.loopingcall [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.934742] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 913.934967] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-898fd202-3018-4753-9acd-5539a1066076 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.960067] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.960067] env[61768]: value = "task-1229196" [ 913.960067] env[61768]: _type = "Task" [ 913.960067] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.969330] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229196, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.994557] env[61768]: DEBUG nova.network.neutron [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Updating instance_info_cache with network_info: [{"id": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "address": "fa:16:3e:a7:3a:6c", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc038ba6f-c7", "ovs_interfaceid": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.043542] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "interface-5cc705a5-1423-43e1-8356-9abd67a8da3a-20999117-9aaf-4291-9e72-f8304dfdf3f0" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.043925] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-5cc705a5-1423-43e1-8356-9abd67a8da3a-20999117-9aaf-4291-9e72-f8304dfdf3f0" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.247187] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.401345] env[61768]: INFO nova.scheduler.client.report [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted allocation for migration 68c66c79-2d4b-4e76-a092-90d11ed31f0b [ 914.477787] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229196, 'name': CreateVM_Task, 'duration_secs': 0.403626} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.477787] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 914.477787] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'delete_on_termination': True, 'device_type': None, 'guest_format': None, 'mount_device': '/dev/sda', 'attachment_id': '16676694-049d-496b-9fc4-f84cdd4f26da', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265498', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'name': 'volume-68c9b861-d548-4ed9-b618-463dfbd8cc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '30f55051-63b3-47a9-83fb-945cdeb82574', 'attached_at': '', 'detached_at': '', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'serial': '68c9b861-d548-4ed9-b618-463dfbd8cc16'}, 'disk_bus': None, 'boot_index': 0, 'volume_type': None}], 'swap': None} {{(pid=61768) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 914.477787] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Root volume attach. Driver type: vmdk {{(pid=61768) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 914.481312] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8664a7db-7b7d-453b-be2e-d929d7e46312 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.490761] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5f0df6f-1abb-4bf2-b600-18936f9a7ba3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.497154] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.499551] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d6571f-3925-4ba9-8197-0ad839902761 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.508022] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-d78f0898-38f9-4bce-8953-619088362565 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.516465] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 914.516465] env[61768]: value = "task-1229197" [ 914.516465] env[61768]: _type = "Task" [ 914.516465] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.534320] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229197, 'name': RelocateVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.547492] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.547602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.550671] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ec2308-f069-46fa-ab00-992b566c4916 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.569921] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ecf0d83-398f-4bf8-9f33-622da6c48996 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.599745] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Reconfiguring VM to detach interface {{(pid=61768) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 914.600129] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2e76c3f-6ae3-4331-be4b-e2a73ca1b81c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.620128] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 914.620128] env[61768]: value = "task-1229198" [ 914.620128] env[61768]: _type = "Task" [ 914.620128] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.628846] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.841367] env[61768]: DEBUG nova.compute.manager [req-21a20394-e020-43ab-81cb-aee85c277862 req-fda1d9e4-bfeb-4231-b656-31f99381c9af service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Received event network-vif-unplugged-c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 914.841677] env[61768]: DEBUG oslo_concurrency.lockutils [req-21a20394-e020-43ab-81cb-aee85c277862 req-fda1d9e4-bfeb-4231-b656-31f99381c9af service nova] Acquiring lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 914.841793] env[61768]: DEBUG oslo_concurrency.lockutils [req-21a20394-e020-43ab-81cb-aee85c277862 req-fda1d9e4-bfeb-4231-b656-31f99381c9af service nova] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 914.841962] env[61768]: DEBUG oslo_concurrency.lockutils [req-21a20394-e020-43ab-81cb-aee85c277862 req-fda1d9e4-bfeb-4231-b656-31f99381c9af service nova] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.842279] env[61768]: DEBUG nova.compute.manager [req-21a20394-e020-43ab-81cb-aee85c277862 req-fda1d9e4-bfeb-4231-b656-31f99381c9af service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] No waiting events found dispatching network-vif-unplugged-c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 914.842389] env[61768]: WARNING nova.compute.manager [req-21a20394-e020-43ab-81cb-aee85c277862 req-fda1d9e4-bfeb-4231-b656-31f99381c9af service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Received unexpected event network-vif-unplugged-c038ba6f-c7aa-47fc-ad80-83bc55817296 for instance with vm_state shelved and task_state shelving_offloading. [ 914.875021] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 914.875021] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f981af-9d78-44f7-b46b-62f931b90ef9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.888320] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 914.888320] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9bcc3d6c-3d1e-4a48-a57f-278b06e02b07 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.907671] env[61768]: DEBUG oslo_concurrency.lockutils [None req-af97254b-2664-4db0-857c-2addd4b52e22 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.552s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.969608] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 914.970192] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 914.970311] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleting the datastore file [datastore2] 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.970648] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fb6a4f8f-d490-4c61-8a58-d74405b93358 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.995439] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 914.995439] env[61768]: value = "task-1229200" [ 914.995439] env[61768]: _type = "Task" [ 914.995439] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.005776] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229200, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.030853] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229197, 'name': RelocateVM_Task, 'duration_secs': 0.428626} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.031546] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 915.031647] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265498', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'name': 'volume-68c9b861-d548-4ed9-b618-463dfbd8cc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '30f55051-63b3-47a9-83fb-945cdeb82574', 'attached_at': '', 'detached_at': '', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'serial': '68c9b861-d548-4ed9-b618-463dfbd8cc16'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 915.032446] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef9ac96-621f-42af-b0c4-f208f4c3e205 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.056623] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e58270-8bf4-4e25-bb61-5ff308c88ed5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.060354] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f991967-8d32-4c3c-8e9e-b09949700373 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.080838] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2e81a7-2df6-451c-ac20-7e5e5f3d7f4d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.093681] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] volume-68c9b861-d548-4ed9-b618-463dfbd8cc16/volume-68c9b861-d548-4ed9-b618-463dfbd8cc16.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.094387] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34266b5c-011e-4288-be36-2350426b0c44 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.158591] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d012dc-ebc8-478f-bf1e-d7abd0c3ea5b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.163376] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 915.163376] env[61768]: value = "task-1229201" [ 915.163376] env[61768]: _type = "Task" [ 915.163376] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.179875] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.182885] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5ae151-f425-45f3-9aa5-a14e454c55e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.195832] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229201, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.213852] env[61768]: DEBUG nova.compute.provider_tree [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.511017] env[61768]: DEBUG oslo_vmware.api [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229200, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.179913} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.511017] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.511017] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 915.511664] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 915.517500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.517500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.517500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.517500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.517500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.522263] env[61768]: INFO nova.compute.manager [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Terminating instance [ 915.524408] env[61768]: DEBUG nova.compute.manager [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 915.524408] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 915.525414] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91cdef99-560c-48d1-9dfa-fcf8c4fa24f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.536304] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 915.536304] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a46756be-0b6b-4882-92d8-eaec5cf36e2e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.547368] env[61768]: DEBUG oslo_vmware.api [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 915.547368] env[61768]: value = "task-1229202" [ 915.547368] env[61768]: _type = "Task" [ 915.547368] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.550588] env[61768]: INFO nova.scheduler.client.report [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted allocations for instance 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015 [ 915.564173] env[61768]: DEBUG oslo_vmware.api [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229202, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.585599] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 915.586531] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 915.586531] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 915.587032] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Rebuilding the list of instances to heal {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 915.667840] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.682898] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229201, 'name': ReconfigVM_Task, 'duration_secs': 0.545546} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.683400] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Reconfigured VM instance instance-0000005b to attach disk [datastore1] volume-68c9b861-d548-4ed9-b618-463dfbd8cc16/volume-68c9b861-d548-4ed9-b618-463dfbd8cc16.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 915.691841] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9839232-b04b-4c70-ad99-788abefca6a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.724513] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 915.724513] env[61768]: value = "task-1229203" [ 915.724513] env[61768]: _type = "Task" [ 915.724513] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.739139] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229203, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.742475] env[61768]: ERROR nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [req-7a6856a1-1c4a-48d8-89e2-21d29931d63e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-7a6856a1-1c4a-48d8-89e2-21d29931d63e"}]} [ 915.766893] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 915.786033] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 915.786300] env[61768]: DEBUG nova.compute.provider_tree [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 915.802224] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 915.825253] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 916.003331] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ac6e17-0977-4b1a-9076-2c1f2e8198d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.016257] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-348a3d34-6385-4458-84c3-31beed093de3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.074945] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.080680] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c1b88c-5188-4e48-842b-fe69582c092b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.095807] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Skipping network cache update for instance because it is being deleted. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 916.096120] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Skipping network cache update for instance because it is being deleted. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 916.096368] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 916.096580] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Skipping network cache update for instance because it is Building. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 916.100251] env[61768]: DEBUG oslo_vmware.api [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229202, 'name': PowerOffVM_Task, 'duration_secs': 0.24751} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.103606] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 916.103930] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 916.104846] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9bb878eb-a136-44a6-9ee3-096f51cf4210 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.108838] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455b6323-a6a7-43ab-b2af-7526f7bc75c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.133453] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquiring lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.133869] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.134219] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquiring lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.134525] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.134901] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.138827] env[61768]: DEBUG nova.compute.provider_tree [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 916.139625] env[61768]: INFO nova.compute.manager [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Terminating instance [ 916.144947] env[61768]: DEBUG nova.compute.manager [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 916.145454] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e577227e-1107-48b8-91b9-d8e988640216 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.164380] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb32b6ac-6687-41a9-979d-17c54a1f24c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.179417] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.206815] env[61768]: WARNING nova.virt.vmwareapi.driver [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e could not be found. [ 916.207052] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 916.209550] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f9e0f09e-eaa1-4960-9d3b-7d5f01fdadb4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.211741] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 916.211999] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 916.212281] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleting the datastore file [datastore2] e2b7cb22-475b-4b1b-b30e-6061e888691e {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 916.216163] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fa33d8ca-73bd-49ad-9c3a-98720e2d0408 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.218447] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "fb344f81-9295-44a1-9bac-dd530c157e98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.218650] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "fb344f81-9295-44a1-9bac-dd530c157e98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.234266] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a356611-a180-4690-9766-4b85dd8139f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.244954] env[61768]: DEBUG oslo_vmware.api [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 916.244954] env[61768]: value = "task-1229205" [ 916.244954] env[61768]: _type = "Task" [ 916.244954] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.245941] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.246098] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquired lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.246301] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Forcefully refreshing network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 916.246444] env[61768]: DEBUG nova.objects.instance [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lazy-loading 'info_cache' on Instance uuid 62cddff5-d499-4ef3-869d-3cdb0328640d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 916.256382] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229203, 'name': ReconfigVM_Task, 'duration_secs': 0.227777} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.257164] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265498', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'name': 'volume-68c9b861-d548-4ed9-b618-463dfbd8cc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '30f55051-63b3-47a9-83fb-945cdeb82574', 'attached_at': '', 'detached_at': '', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'serial': '68c9b861-d548-4ed9-b618-463dfbd8cc16'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 916.258020] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa6d1527-e41e-4af1-a434-e7fce24fc925 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.263617] env[61768]: DEBUG oslo_vmware.api [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229205, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.276943] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e could not be found. [ 916.277179] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.277377] env[61768]: INFO nova.compute.manager [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Took 0.13 seconds to destroy the instance on the hypervisor. [ 916.277681] env[61768]: DEBUG oslo.service.loopingcall [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.279486] env[61768]: DEBUG nova.compute.manager [-] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.279591] env[61768]: DEBUG nova.network.neutron [-] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 916.281508] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 916.281508] env[61768]: value = "task-1229206" [ 916.281508] env[61768]: _type = "Task" [ 916.281508] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.291787] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229206, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.664860] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.674881] env[61768]: ERROR nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [req-00f71fcc-61ea-4d16-866e-78072c15645f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f9f7149c-7c4e-44e7-8000-07a202159d2f. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-00f71fcc-61ea-4d16-866e-78072c15645f"}]} [ 916.696075] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Refreshing inventories for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 916.714692] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating ProviderTree inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 916.714936] env[61768]: DEBUG nova.compute.provider_tree [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 150, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 916.721672] env[61768]: DEBUG nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 916.733918] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Refreshing aggregate associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, aggregates: None {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 916.762817] env[61768]: DEBUG oslo_vmware.api [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229205, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.414441} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.764832] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Refreshing trait associations for resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=61768) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 916.766966] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 916.767260] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 916.767361] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 916.767536] env[61768]: INFO nova.compute.manager [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Took 1.24 seconds to destroy the instance on the hypervisor. [ 916.767803] env[61768]: DEBUG oslo.service.loopingcall [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 916.768467] env[61768]: DEBUG nova.compute.manager [-] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 916.768568] env[61768]: DEBUG nova.network.neutron [-] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 916.792619] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229206, 'name': Rename_Task, 'duration_secs': 0.217532} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.792954] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 916.793231] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6a37c29b-796e-40c9-9ed4-35b472b301ec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.804560] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 916.804560] env[61768]: value = "task-1229207" [ 916.804560] env[61768]: _type = "Task" [ 916.804560] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.814661] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.960880] env[61768]: DEBUG nova.compute.manager [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Received event network-changed-c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 916.961122] env[61768]: DEBUG nova.compute.manager [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Refreshing instance network info cache due to event network-changed-c038ba6f-c7aa-47fc-ad80-83bc55817296. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 916.961364] env[61768]: DEBUG oslo_concurrency.lockutils [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] Acquiring lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.961536] env[61768]: DEBUG oslo_concurrency.lockutils [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] Acquired lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.961694] env[61768]: DEBUG nova.network.neutron [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Refreshing network info cache for port c038ba6f-c7aa-47fc-ad80-83bc55817296 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 916.989212] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b58f160-a184-4662-b03b-8856eb8e2148 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.001424] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2c1366-f43a-43b7-93ce-ad758cdea54e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.041471] env[61768]: DEBUG oslo_concurrency.lockutils [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.042043] env[61768]: DEBUG nova.network.neutron [-] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.043618] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-076cd9e8-b839-48b2-9401-e8dd0b78c921 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.048395] env[61768]: DEBUG nova.compute.manager [req-6384b8d6-8488-45fa-8d08-f0054bc9161a req-6b96ed96-6ebd-49a9-bb76-3773df429572 service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Received event network-vif-deleted-5b4c6981-f46a-4e25-9f9a-51106a3988ac {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.048610] env[61768]: INFO nova.compute.manager [req-6384b8d6-8488-45fa-8d08-f0054bc9161a req-6b96ed96-6ebd-49a9-bb76-3773df429572 service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Neutron deleted interface 5b4c6981-f46a-4e25-9f9a-51106a3988ac; detaching it from the instance and deleting it from the info cache [ 917.048783] env[61768]: DEBUG nova.network.neutron [req-6384b8d6-8488-45fa-8d08-f0054bc9161a req-6b96ed96-6ebd-49a9-bb76-3773df429572 service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.059026] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b194791-492f-4e68-84d4-b06ee19e9599 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.073153] env[61768]: DEBUG nova.compute.provider_tree [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.163535] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.250868] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.321714] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229207, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.548975] env[61768]: INFO nova.compute.manager [-] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Took 0.78 seconds to deallocate network for instance. [ 917.557866] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e50bb8f6-5a62-4c20-b505-e2b3441472d0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.574266] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c898d54d-0141-43d8-ae88-fe91728780f3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.623087] env[61768]: DEBUG nova.compute.manager [req-6384b8d6-8488-45fa-8d08-f0054bc9161a req-6b96ed96-6ebd-49a9-bb76-3773df429572 service nova] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Detach interface failed, port_id=5b4c6981-f46a-4e25-9f9a-51106a3988ac, reason: Instance e2b7cb22-475b-4b1b-b30e-6061e888691e could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 917.635018] env[61768]: DEBUG nova.scheduler.client.report [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updated inventory for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with generation 114 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 917.635018] env[61768]: DEBUG nova.compute.provider_tree [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating resource provider f9f7149c-7c4e-44e7-8000-07a202159d2f generation from 114 to 115 during operation: update_inventory {{(pid=61768) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 917.635018] env[61768]: DEBUG nova.compute.provider_tree [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Updating inventory in ProviderTree for provider f9f7149c-7c4e-44e7-8000-07a202159d2f with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.664261] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.690439] env[61768]: DEBUG nova.network.neutron [-] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.705505] env[61768]: DEBUG nova.network.neutron [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Updated VIF entry in instance network info cache for port c038ba6f-c7aa-47fc-ad80-83bc55817296. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.705961] env[61768]: DEBUG nova.network.neutron [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Updating instance_info_cache with network_info: [{"id": "c038ba6f-c7aa-47fc-ad80-83bc55817296", "address": "fa:16:3e:a7:3a:6c", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": null, "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapc038ba6f-c7", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.817901] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229207, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.012752] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [{"id": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "address": "fa:16:3e:fc:98:a8", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap44b2acbd-7a", "ovs_interfaceid": "44b2acbd-7a9f-416b-b3b4-e877de5bfa85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.062432] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.138031] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.306s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.138576] env[61768]: DEBUG nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 918.141893] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.894s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.168258] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.194295] env[61768]: INFO nova.compute.manager [-] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Took 1.91 seconds to deallocate network for instance. [ 918.211219] env[61768]: DEBUG oslo_concurrency.lockutils [req-e383722e-4c7b-4e7c-829e-f13aaea9df56 req-421cc07e-c557-444a-a1d1-7f30abf164a5 service nova] Releasing lock "refresh_cache-4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.327648] env[61768]: DEBUG oslo_vmware.api [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229207, 'name': PowerOnVM_Task, 'duration_secs': 1.298752} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.328227] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 918.328516] env[61768]: INFO nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Took 6.77 seconds to spawn the instance on the hypervisor. [ 918.328837] env[61768]: DEBUG nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 918.330836] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04592d5-dac5-4ffc-9e21-f38507d17830 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.346614] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f198e3c9-078d-4d2f-8909-02cc4a21943f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.354839] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57bb93ec-847d-4f2c-848d-0ae2edd6c166 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.393656] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71646a4f-3620-49d7-972f-0a15c302c33d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.402670] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80f4ae83-3456-4c5a-a636-78977c28f1e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.417741] env[61768]: DEBUG nova.compute.provider_tree [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.516970] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Releasing lock "refresh_cache-62cddff5-d499-4ef3-869d-3cdb0328640d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.517477] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updated the network info_cache for instance {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 918.517807] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.518197] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.518527] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.518829] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.519160] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.519419] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.519645] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 918.519932] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.648126] env[61768]: DEBUG nova.compute.utils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.649024] env[61768]: DEBUG nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.649209] env[61768]: DEBUG nova.network.neutron [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.671027] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.741997] env[61768]: DEBUG nova.policy [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fd5852bc16429aa2904b4351c8249d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7857b5217cc499ca8c9c3b8a5471db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.755728] env[61768]: INFO nova.compute.manager [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Took 0.56 seconds to detach 1 volumes for instance. [ 918.758540] env[61768]: DEBUG nova.compute.manager [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Deleting volume: 9ddc3f0e-f5f4-449a-878f-bda8467085c3 {{(pid=61768) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 918.857314] env[61768]: INFO nova.compute.manager [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Took 13.52 seconds to build instance. [ 918.920927] env[61768]: DEBUG nova.scheduler.client.report [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 918.992476] env[61768]: DEBUG nova.compute.manager [req-d7d5f426-6bd4-4265-aad5-b298e4e90bbd req-fa2be416-a985-4aef-85ee-57438758dae2 service nova] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Received event network-vif-deleted-aab2c242-227c-4cf0-9a64-75e0237360d7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.023507] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.154885] env[61768]: DEBUG nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 919.173241] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.307222] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.318815] env[61768]: DEBUG nova.network.neutron [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Successfully created port: 5b0fd366-9b1d-4136-8996-73c527e03783 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.361803] env[61768]: DEBUG oslo_concurrency.lockutils [None req-21e69d55-e495-4567-88d7-fbb30d87e0e8 tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "30f55051-63b3-47a9-83fb-945cdeb82574" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.037s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.427921] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.287s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.431018] env[61768]: INFO nova.compute.manager [None req-9d1fd6b7-fe0e-4840-9d71-45e22c088ce5 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Successfully reverted task state from rebuilding on failure for instance. [ 919.441025] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.363s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.441025] env[61768]: DEBUG nova.objects.instance [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'resources' on Instance uuid 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 919.680906] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.943505] env[61768]: DEBUG nova.objects.instance [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'numa_topology' on Instance uuid 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 920.170609] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.177257] env[61768]: DEBUG nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 920.210908] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 920.211199] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 920.211363] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 920.211642] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 920.211707] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 920.211861] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 920.212096] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 920.212275] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 920.212458] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 920.212628] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 920.212808] env[61768]: DEBUG nova.virt.hardware [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 920.213741] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a578c399-9e14-4c4b-965e-46bfd56e2d93 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.222885] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcb82793-1ac7-4851-b76e-fdb382fcfc42 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.340787] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "1158fbc3-f17b-44f7-847f-01bdf328a74a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.341063] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.451097] env[61768]: DEBUG nova.objects.base [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Object Instance<4cc5b1fa-55e3-4b2d-b10d-8da7b3356015> lazy-loaded attributes: resources,numa_topology {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 920.672382] env[61768]: DEBUG oslo_vmware.api [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229198, 'name': ReconfigVM_Task, 'duration_secs': 5.798884} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.672703] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.672956] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Reconfigured VM to detach interface {{(pid=61768) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 920.732628] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8210b22b-ecc0-4e10-aed4-881c78993174 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.741164] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-030baddb-cad9-4c90-8507-6e88c1c51405 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.772766] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f442a180-3bba-456f-ba49-638d0ec72946 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.781873] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e4f89d8-93e2-4223-8bd7-f107d5886536 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.799266] env[61768]: DEBUG nova.compute.provider_tree [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 920.847074] env[61768]: DEBUG nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 921.027449] env[61768]: DEBUG nova.compute.manager [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Received event network-changed-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.033127] env[61768]: DEBUG nova.compute.manager [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Refreshing instance network info cache due to event network-changed-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.033127] env[61768]: DEBUG oslo_concurrency.lockutils [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] Acquiring lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.033127] env[61768]: DEBUG oslo_concurrency.lockutils [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] Acquired lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.033127] env[61768]: DEBUG nova.network.neutron [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Refreshing network info cache for port 4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 921.302771] env[61768]: DEBUG nova.scheduler.client.report [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 921.372583] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.486705] env[61768]: DEBUG nova.network.neutron [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Successfully updated port: 5b0fd366-9b1d-4136-8996-73c527e03783 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 921.745775] env[61768]: DEBUG nova.network.neutron [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Updated VIF entry in instance network info cache for port 4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.746183] env[61768]: DEBUG nova.network.neutron [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Updating instance_info_cache with network_info: [{"id": "4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3", "address": "fa:16:3e:f3:d1:42", "network": {"id": "24ffcec7-794b-4e30-984c-4ccb9657d666", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-301275445-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.205", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f958bb192d0349c4ac4f256db207d575", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cd098b1c-636f-492d-b5ae-037cb0cae454", "external-id": "nsx-vlan-transportzone-377", "segmentation_id": 377, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4cd7e358-0e", "ovs_interfaceid": "4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.807595] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.370s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.817897] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.560s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.817897] env[61768]: INFO nova.compute.claims [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 921.990672] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "refresh_cache-e0c89fa6-9d32-4343-89d7-e9129d252b40" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.990856] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "refresh_cache-e0c89fa6-9d32-4343-89d7-e9129d252b40" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.991045] env[61768]: DEBUG nova.network.neutron [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.248469] env[61768]: DEBUG oslo_concurrency.lockutils [req-ad12df88-904f-4aae-a00d-89845b306089 req-c755684a-6d89-4887-917a-06fd0368564a service nova] Releasing lock "refresh_cache-30f55051-63b3-47a9-83fb-945cdeb82574" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.276292] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.276491] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquired lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.276677] env[61768]: DEBUG nova.network.neutron [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 922.330464] env[61768]: DEBUG oslo_concurrency.lockutils [None req-eacfb90d-39d5-4067-aa94-64f3a254e42e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 26.322s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.333101] env[61768]: DEBUG oslo_concurrency.lockutils [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.291s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.333101] env[61768]: DEBUG oslo_concurrency.lockutils [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.333250] env[61768]: DEBUG oslo_concurrency.lockutils [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.333316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.340292] env[61768]: INFO nova.compute.manager [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Terminating instance [ 922.342304] env[61768]: DEBUG nova.compute.manager [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.342304] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.342304] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ba696eae-b4ab-4741-9261-5d02a6768c33 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.353598] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c010e40b-fc26-4fd7-a386-169f54ef4ab1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.387753] env[61768]: WARNING nova.virt.vmwareapi.vmops [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015 could not be found. [ 922.388096] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 922.388386] env[61768]: INFO nova.compute.manager [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Took 0.05 seconds to destroy the instance on the hypervisor. [ 922.388878] env[61768]: DEBUG oslo.service.loopingcall [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.389223] env[61768]: DEBUG nova.compute.manager [-] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 922.389379] env[61768]: DEBUG nova.network.neutron [-] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 922.522872] env[61768]: DEBUG nova.network.neutron [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 922.673181] env[61768]: DEBUG nova.network.neutron [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Updating instance_info_cache with network_info: [{"id": "5b0fd366-9b1d-4136-8996-73c527e03783", "address": "fa:16:3e:8f:26:ea", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b0fd366-9b", "ovs_interfaceid": "5b0fd366-9b1d-4136-8996-73c527e03783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.748225] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.748804] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.749041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.749244] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.749583] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.751701] env[61768]: INFO nova.compute.manager [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Terminating instance [ 922.753911] env[61768]: DEBUG nova.compute.manager [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 922.755320] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 922.759658] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-927d1225-b5c9-4ea3-9aa9-50017ce48f86 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.767859] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 922.767859] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d0795b4-a995-4e19-bbe9-9a6f3e3e04f3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.774859] env[61768]: DEBUG oslo_vmware.api [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 922.774859] env[61768]: value = "task-1229209" [ 922.774859] env[61768]: _type = "Task" [ 922.774859] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.787659] env[61768]: DEBUG oslo_vmware.api [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.018927] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a3dbfe-80e1-4650-9736-4305765df63f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.030833] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a503f84c-38c8-4bc8-b54f-67629400ad80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.074462] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c5b49a4-7560-48c4-aedf-af4f0a6f5a97 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.082959] env[61768]: DEBUG nova.compute.manager [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Received event network-vif-plugged-5b0fd366-9b1d-4136-8996-73c527e03783 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.082959] env[61768]: DEBUG oslo_concurrency.lockutils [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] Acquiring lock "e0c89fa6-9d32-4343-89d7-e9129d252b40-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.082959] env[61768]: DEBUG oslo_concurrency.lockutils [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.082959] env[61768]: DEBUG oslo_concurrency.lockutils [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.082959] env[61768]: DEBUG nova.compute.manager [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] No waiting events found dispatching network-vif-plugged-5b0fd366-9b1d-4136-8996-73c527e03783 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.082959] env[61768]: WARNING nova.compute.manager [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Received unexpected event network-vif-plugged-5b0fd366-9b1d-4136-8996-73c527e03783 for instance with vm_state building and task_state spawning. [ 923.082959] env[61768]: DEBUG nova.compute.manager [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Received event network-changed-5b0fd366-9b1d-4136-8996-73c527e03783 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 923.082959] env[61768]: DEBUG nova.compute.manager [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Refreshing instance network info cache due to event network-changed-5b0fd366-9b1d-4136-8996-73c527e03783. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 923.082959] env[61768]: DEBUG oslo_concurrency.lockutils [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] Acquiring lock "refresh_cache-e0c89fa6-9d32-4343-89d7-e9129d252b40" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.087188] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f55b02-2eb4-47ac-9bbe-8fe5b962f36b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.106293] env[61768]: DEBUG nova.compute.provider_tree [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.176174] env[61768]: INFO nova.network.neutron [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Port 20999117-9aaf-4291-9e72-f8304dfdf3f0 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 923.176603] env[61768]: DEBUG nova.network.neutron [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [{"id": "604a3820-78ee-48e4-bb9e-af1273963ff8", "address": "fa:16:3e:f2:65:3e", "network": {"id": "2fbe88f9-1bca-4347-914d-a62dc6bebde8", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1722742721-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.228", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2f3e8b04445948a6a8a737ad89a5c002", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d8383707-f093-40a7-a5ba-31b0e07cac45", "external-id": "cl2-zone-18", "segmentation_id": 18, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap604a3820-78", "ovs_interfaceid": "604a3820-78ee-48e4-bb9e-af1273963ff8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.178275] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "refresh_cache-e0c89fa6-9d32-4343-89d7-e9129d252b40" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.178275] env[61768]: DEBUG nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Instance network_info: |[{"id": "5b0fd366-9b1d-4136-8996-73c527e03783", "address": "fa:16:3e:8f:26:ea", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b0fd366-9b", "ovs_interfaceid": "5b0fd366-9b1d-4136-8996-73c527e03783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 923.178730] env[61768]: DEBUG oslo_concurrency.lockutils [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] Acquired lock "refresh_cache-e0c89fa6-9d32-4343-89d7-e9129d252b40" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.179362] env[61768]: DEBUG nova.network.neutron [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Refreshing network info cache for port 5b0fd366-9b1d-4136-8996-73c527e03783 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 923.181857] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:26:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b0fd366-9b1d-4136-8996-73c527e03783', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 923.187295] env[61768]: DEBUG oslo.service.loopingcall [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.189854] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 923.189854] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d503eec-2143-423f-b1f2-bbf866dcd9ca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.211268] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 923.211268] env[61768]: value = "task-1229210" [ 923.211268] env[61768]: _type = "Task" [ 923.211268] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.220969] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229210, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.263633] env[61768]: DEBUG nova.network.neutron [-] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.286325] env[61768]: DEBUG oslo_vmware.api [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229209, 'name': PowerOffVM_Task, 'duration_secs': 0.38935} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.287536] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 923.287536] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 923.288495] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61d2b67b-baac-4ac7-bd1d-d44fb5c6f56a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.368997] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 923.371170] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 923.371170] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleting the datastore file [datastore2] 5cc705a5-1423-43e1-8356-9abd67a8da3a {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 923.371170] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e538d128-fe47-4857-8577-18c1085c0e01 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.377873] env[61768]: DEBUG oslo_vmware.api [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 923.377873] env[61768]: value = "task-1229212" [ 923.377873] env[61768]: _type = "Task" [ 923.377873] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.386671] env[61768]: DEBUG oslo_vmware.api [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229212, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.609091] env[61768]: DEBUG nova.scheduler.client.report [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 923.680019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Releasing lock "refresh_cache-5cc705a5-1423-43e1-8356-9abd67a8da3a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.722281] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229210, 'name': CreateVM_Task, 'duration_secs': 0.336277} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.722768] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 923.723468] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.723640] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.723972] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 923.724322] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7eaabafe-fed3-41db-b975-45cb8e034259 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.729775] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 923.729775] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bff72c-891d-b681-1fb0-c131a7e5607d" [ 923.729775] env[61768]: _type = "Task" [ 923.729775] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.738930] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bff72c-891d-b681-1fb0-c131a7e5607d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.768493] env[61768]: INFO nova.compute.manager [-] [instance: 4cc5b1fa-55e3-4b2d-b10d-8da7b3356015] Took 1.38 seconds to deallocate network for instance. [ 923.890063] env[61768]: DEBUG oslo_vmware.api [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229212, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.168648} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.890341] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 923.890502] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 923.890686] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 923.890874] env[61768]: INFO nova.compute.manager [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Took 1.14 seconds to destroy the instance on the hypervisor. [ 923.891192] env[61768]: DEBUG oslo.service.loopingcall [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 923.891487] env[61768]: DEBUG nova.compute.manager [-] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 923.891487] env[61768]: DEBUG nova.network.neutron [-] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 923.988380] env[61768]: DEBUG nova.network.neutron [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Updated VIF entry in instance network info cache for port 5b0fd366-9b1d-4136-8996-73c527e03783. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.988959] env[61768]: DEBUG nova.network.neutron [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Updating instance_info_cache with network_info: [{"id": "5b0fd366-9b1d-4136-8996-73c527e03783", "address": "fa:16:3e:8f:26:ea", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b0fd366-9b", "ovs_interfaceid": "5b0fd366-9b1d-4136-8996-73c527e03783", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.114253] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.114959] env[61768]: DEBUG nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 924.118138] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.056s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.118538] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.120792] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.097s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.120952] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.121162] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 924.121502] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.815s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.121778] env[61768]: DEBUG nova.objects.instance [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lazy-loading 'resources' on Instance uuid c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 924.125080] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68498bc0-6631-413c-9321-e5896dc982c9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.143088] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f9cdcd6-a277-454f-8551-128486c59300 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.166013] env[61768]: INFO nova.scheduler.client.report [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted allocations for instance e2b7cb22-475b-4b1b-b30e-6061e888691e [ 924.168685] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a86c3a-3d22-4ced-b8e2-4688acbc4381 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.182019] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a786b6-4fa3-466d-a869-db3d6ca4c9df {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.184021] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6bc14483-8b41-4b9b-8a69-a183b15671ea tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "interface-5cc705a5-1423-43e1-8356-9abd67a8da3a-20999117-9aaf-4291-9e72-f8304dfdf3f0" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.140s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.213037] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179800MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 924.213185] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.241696] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52bff72c-891d-b681-1fb0-c131a7e5607d, 'name': SearchDatastore_Task, 'duration_secs': 0.009654} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.242127] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.242423] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 924.242668] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.242817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.242990] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 924.243258] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-89ef83eb-4c3a-4680-b695-6b034ade34fc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.253649] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 924.253806] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 924.254554] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad4596e0-064b-49f1-bc19-7357317671d2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.260269] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 924.260269] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5281415a-bfa3-c3a3-a2d2-b39c5d2f3232" [ 924.260269] env[61768]: _type = "Task" [ 924.260269] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.270208] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5281415a-bfa3-c3a3-a2d2-b39c5d2f3232, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.492160] env[61768]: DEBUG oslo_concurrency.lockutils [req-01e55cda-42a5-49e4-a118-eb5e3becc5ba req-6e5a4a51-cccc-4b68-b4fb-8ef47fdb4728 service nova] Releasing lock "refresh_cache-e0c89fa6-9d32-4343-89d7-e9129d252b40" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.627057] env[61768]: DEBUG nova.compute.utils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 924.627057] env[61768]: DEBUG nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 924.627057] env[61768]: DEBUG nova.network.neutron [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 924.641176] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "df570987-0144-4e3a-83db-a9538d3835f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.641176] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "df570987-0144-4e3a-83db-a9538d3835f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.679183] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a7f3ba98-cb6e-49bb-ac2a-c86f1a8a206b tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "e2b7cb22-475b-4b1b-b30e-6061e888691e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.162s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.716154] env[61768]: DEBUG nova.policy [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a9d7f72b4c22452d89004ca70d165e67', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f4804c086b449608ef34ea22c018907', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.778319] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5281415a-bfa3-c3a3-a2d2-b39c5d2f3232, 'name': SearchDatastore_Task, 'duration_secs': 0.012976} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.779215] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a64e7d2-4344-40ee-b625-bd89828dabc0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.785977] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 924.785977] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c1f4c9-7ce7-0b35-b0a0-401ac6445cc4" [ 924.785977] env[61768]: _type = "Task" [ 924.785977] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.797631] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c1f4c9-7ce7-0b35-b0a0-401ac6445cc4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.798230] env[61768]: DEBUG oslo_concurrency.lockutils [None req-10d0b75a-ce62-4e16-80ad-569fece3bec7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "4cc5b1fa-55e3-4b2d-b10d-8da7b3356015" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.465s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.859036] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9391722-fe71-487d-9144-5f6cb3e75db3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.861124] env[61768]: DEBUG nova.compute.manager [req-d2ef4605-8bbc-4a89-9e03-80b6675032dd req-1394edf9-7ec4-44ad-99a8-f4c5e30add1d service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Received event network-vif-deleted-604a3820-78ee-48e4-bb9e-af1273963ff8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 924.861385] env[61768]: INFO nova.compute.manager [req-d2ef4605-8bbc-4a89-9e03-80b6675032dd req-1394edf9-7ec4-44ad-99a8-f4c5e30add1d service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Neutron deleted interface 604a3820-78ee-48e4-bb9e-af1273963ff8; detaching it from the instance and deleting it from the info cache [ 924.861522] env[61768]: DEBUG nova.network.neutron [req-d2ef4605-8bbc-4a89-9e03-80b6675032dd req-1394edf9-7ec4-44ad-99a8-f4c5e30add1d service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.873706] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe007e4c-5e13-4856-87a3-f6aafcff5026 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.916898] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7841575a-d2da-45d9-a63f-53b41296a03f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.925911] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ee05218-38b7-48d0-b30f-50835381b0c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.944342] env[61768]: DEBUG nova.compute.provider_tree [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 925.041485] env[61768]: DEBUG nova.network.neutron [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Successfully created port: 51ac6b12-5631-4ce5-acc4-2f43dee834d0 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 925.131750] env[61768]: DEBUG nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 925.144520] env[61768]: DEBUG nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 925.297028] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c1f4c9-7ce7-0b35-b0a0-401ac6445cc4, 'name': SearchDatastore_Task, 'duration_secs': 0.013236} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.297028] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.297028] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] e0c89fa6-9d32-4343-89d7-e9129d252b40/e0c89fa6-9d32-4343-89d7-e9129d252b40.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 925.297427] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf14136a-305c-4719-adea-26ed8522ee30 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.305069] env[61768]: DEBUG nova.network.neutron [-] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.306508] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 925.306508] env[61768]: value = "task-1229213" [ 925.306508] env[61768]: _type = "Task" [ 925.306508] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.314981] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229213, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.366526] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7aeae03c-4bc0-4401-8b46-27e5ae67388d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.374712] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58534dcb-6910-4e2a-9de3-b60810951e5a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.409533] env[61768]: DEBUG nova.compute.manager [req-d2ef4605-8bbc-4a89-9e03-80b6675032dd req-1394edf9-7ec4-44ad-99a8-f4c5e30add1d service nova] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Detach interface failed, port_id=604a3820-78ee-48e4-bb9e-af1273963ff8, reason: Instance 5cc705a5-1423-43e1-8356-9abd67a8da3a could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 925.412996] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.413258] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.452117] env[61768]: DEBUG nova.scheduler.client.report [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.670782] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.807109] env[61768]: INFO nova.compute.manager [-] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Took 1.92 seconds to deallocate network for instance. [ 925.821741] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229213, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450978} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.822018] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] e0c89fa6-9d32-4343-89d7-e9129d252b40/e0c89fa6-9d32-4343-89d7-e9129d252b40.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 925.822251] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 925.822500] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82987ec6-0e76-4be1-b2eb-13559eaef02c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.829648] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 925.829648] env[61768]: value = "task-1229214" [ 925.829648] env[61768]: _type = "Task" [ 925.829648] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.838723] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229214, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.915975] env[61768]: DEBUG nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 925.957412] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.836s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.960158] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.588s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.961858] env[61768]: INFO nova.compute.claims [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.141529] env[61768]: DEBUG nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 926.167489] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 926.167721] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 926.167915] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 926.168133] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 926.168293] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 926.168449] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 926.168663] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 926.168832] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 926.169016] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 926.169297] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 926.169589] env[61768]: DEBUG nova.virt.hardware [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 926.170527] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd1e6bbf-ce9b-47b6-9b59-70133ac0934a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.179351] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c130e69-45ff-41b6-acf3-a97c61230c9e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.318341] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.339855] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229214, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067375} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.340216] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 926.341096] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2ae1fb-8d5b-4116-beba-b25bb645fb89 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.363436] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Reconfiguring VM instance instance-0000005c to attach disk [datastore2] e0c89fa6-9d32-4343-89d7-e9129d252b40/e0c89fa6-9d32-4343-89d7-e9129d252b40.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 926.363705] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2e0bd49-6826-47e0-9e9f-98ae0e901a80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.385585] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 926.385585] env[61768]: value = "task-1229215" [ 926.385585] env[61768]: _type = "Task" [ 926.385585] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.394118] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229215, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.434409] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.452763] env[61768]: DEBUG nova.compute.manager [req-a0edf351-b69a-4ef2-9cd5-c70bc5549e83 req-8cfb48bb-26e7-4529-801b-79ee6c525fc3 service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Received event network-vif-plugged-51ac6b12-5631-4ce5-acc4-2f43dee834d0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.452994] env[61768]: DEBUG oslo_concurrency.lockutils [req-a0edf351-b69a-4ef2-9cd5-c70bc5549e83 req-8cfb48bb-26e7-4529-801b-79ee6c525fc3 service nova] Acquiring lock "fb344f81-9295-44a1-9bac-dd530c157e98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.453255] env[61768]: DEBUG oslo_concurrency.lockutils [req-a0edf351-b69a-4ef2-9cd5-c70bc5549e83 req-8cfb48bb-26e7-4529-801b-79ee6c525fc3 service nova] Lock "fb344f81-9295-44a1-9bac-dd530c157e98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.453402] env[61768]: DEBUG oslo_concurrency.lockutils [req-a0edf351-b69a-4ef2-9cd5-c70bc5549e83 req-8cfb48bb-26e7-4529-801b-79ee6c525fc3 service nova] Lock "fb344f81-9295-44a1-9bac-dd530c157e98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.453573] env[61768]: DEBUG nova.compute.manager [req-a0edf351-b69a-4ef2-9cd5-c70bc5549e83 req-8cfb48bb-26e7-4529-801b-79ee6c525fc3 service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] No waiting events found dispatching network-vif-plugged-51ac6b12-5631-4ce5-acc4-2f43dee834d0 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.453743] env[61768]: WARNING nova.compute.manager [req-a0edf351-b69a-4ef2-9cd5-c70bc5549e83 req-8cfb48bb-26e7-4529-801b-79ee6c525fc3 service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Received unexpected event network-vif-plugged-51ac6b12-5631-4ce5-acc4-2f43dee834d0 for instance with vm_state building and task_state spawning. [ 926.477852] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f302c418-5775-495c-a676-8f90dda5a462 tempest-ServerActionsV293TestJSON-21745575 tempest-ServerActionsV293TestJSON-21745575-project-member] Lock "c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.344s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.897297] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229215, 'name': ReconfigVM_Task, 'duration_secs': 0.304597} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.897591] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Reconfigured VM instance instance-0000005c to attach disk [datastore2] e0c89fa6-9d32-4343-89d7-e9129d252b40/e0c89fa6-9d32-4343-89d7-e9129d252b40.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 926.898253] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-937b30bd-f0cb-4b99-8873-e948b5715d47 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.906675] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 926.906675] env[61768]: value = "task-1229216" [ 926.906675] env[61768]: _type = "Task" [ 926.906675] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.916168] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229216, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.028967] env[61768]: DEBUG nova.network.neutron [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Successfully updated port: 51ac6b12-5631-4ce5-acc4-2f43dee834d0 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 927.054198] env[61768]: DEBUG nova.compute.manager [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Received event network-changed-51ac6b12-5631-4ce5-acc4-2f43dee834d0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.054433] env[61768]: DEBUG nova.compute.manager [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Refreshing instance network info cache due to event network-changed-51ac6b12-5631-4ce5-acc4-2f43dee834d0. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 927.054845] env[61768]: DEBUG oslo_concurrency.lockutils [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] Acquiring lock "refresh_cache-fb344f81-9295-44a1-9bac-dd530c157e98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.055025] env[61768]: DEBUG oslo_concurrency.lockutils [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] Acquired lock "refresh_cache-fb344f81-9295-44a1-9bac-dd530c157e98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.055205] env[61768]: DEBUG nova.network.neutron [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Refreshing network info cache for port 51ac6b12-5631-4ce5-acc4-2f43dee834d0 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 927.147588] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2119f324-8e6a-415c-8767-af9a2f798e4e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.155963] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd50328-dad6-443a-a94a-17626ac970a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.186393] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef68448-ce69-4c21-8653-495c9895f622 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.196361] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23661395-1313-4b42-9750-272b128e6282 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.210318] env[61768]: DEBUG nova.compute.provider_tree [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 927.423383] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229216, 'name': Rename_Task, 'duration_secs': 0.149204} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.423818] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 927.424194] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4a869b45-ac6a-44c6-8f9f-0ecac5424405 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.431597] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 927.431597] env[61768]: value = "task-1229217" [ 927.431597] env[61768]: _type = "Task" [ 927.431597] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.442842] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229217, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.531878] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "refresh_cache-fb344f81-9295-44a1-9bac-dd530c157e98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.589760] env[61768]: DEBUG nova.network.neutron [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.697125] env[61768]: DEBUG nova.network.neutron [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.713604] env[61768]: DEBUG nova.scheduler.client.report [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.942976] env[61768]: DEBUG oslo_vmware.api [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229217, 'name': PowerOnVM_Task, 'duration_secs': 0.449614} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.943106] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 927.943319] env[61768]: INFO nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Took 7.77 seconds to spawn the instance on the hypervisor. [ 927.943509] env[61768]: DEBUG nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.944615] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d906581-fbbe-4121-85fd-7a8026d75c67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.200985] env[61768]: DEBUG oslo_concurrency.lockutils [req-b8085e3b-a588-476e-9f38-204340615f5e req-a8d984a4-3ad7-4eb2-a956-5101e63e4fac service nova] Releasing lock "refresh_cache-fb344f81-9295-44a1-9bac-dd530c157e98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.200985] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "refresh_cache-fb344f81-9295-44a1-9bac-dd530c157e98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.201352] env[61768]: DEBUG nova.network.neutron [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.217967] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.258s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.218690] env[61768]: DEBUG nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 928.221662] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 4.008s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.461409] env[61768]: INFO nova.compute.manager [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Took 17.96 seconds to build instance. [ 928.724639] env[61768]: DEBUG nova.compute.utils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.731619] env[61768]: DEBUG nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 928.731619] env[61768]: DEBUG nova.network.neutron [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 928.734606] env[61768]: DEBUG nova.network.neutron [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 928.777480] env[61768]: DEBUG nova.policy [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a45b47835a87452fbded894a108803a2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '64f5133fd59844859b8e01d31f3320ab', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 928.924980] env[61768]: DEBUG nova.network.neutron [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Updating instance_info_cache with network_info: [{"id": "51ac6b12-5631-4ce5-acc4-2f43dee834d0", "address": "fa:16:3e:82:95:dd", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51ac6b12-56", "ovs_interfaceid": "51ac6b12-5631-4ce5-acc4-2f43dee834d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.963603] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0328548f-18e2-421b-b09d-35883134cf07 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.470s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.057529] env[61768]: DEBUG nova.network.neutron [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Successfully created port: edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 929.062112] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "e0c89fa6-9d32-4343-89d7-e9129d252b40" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.062395] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.062615] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "e0c89fa6-9d32-4343-89d7-e9129d252b40-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.062803] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.062982] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.065352] env[61768]: INFO nova.compute.manager [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Terminating instance [ 929.067516] env[61768]: DEBUG nova.compute.manager [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 929.068197] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 929.068595] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f09d0e3-e72a-4ff2-95ce-2711f6353f3f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.079322] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 929.079679] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2de7e1e2-ba10-4f9b-98b3-fc453d47e79b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.087893] env[61768]: DEBUG oslo_vmware.api [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 929.087893] env[61768]: value = "task-1229218" [ 929.087893] env[61768]: _type = "Task" [ 929.087893] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.101661] env[61768]: DEBUG oslo_vmware.api [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229218, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.238044] env[61768]: DEBUG nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 929.267928] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 2c85de5d-0111-40a6-a8bc-69c7eba0393d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.268252] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 7697ecda-ef30-447d-a92a-3f5cb3cc9118 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.268631] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 8e0e13a7-eac5-4176-8536-2906d13c390e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.268810] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance ecd08348-b22f-49c5-b13f-b187506e38d6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.271024] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 62cddff5-d499-4ef3-869d-3cdb0328640d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.271024] env[61768]: WARNING nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 5cc705a5-1423-43e1-8356-9abd67a8da3a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 929.271024] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance d54c94a6-0a53-47a8-b024-1a05439da837 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.271024] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 739faae0-5811-4f88-b56b-1350c0f7b8be actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.271024] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 30f55051-63b3-47a9-83fb-945cdeb82574 actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.271024] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance e0c89fa6-9d32-4343-89d7-e9129d252b40 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.271024] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance fb344f81-9295-44a1-9bac-dd530c157e98 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 929.271024] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 1158fbc3-f17b-44f7-847f-01bdf328a74a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 930.036849] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "refresh_cache-fb344f81-9295-44a1-9bac-dd530c157e98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.037283] env[61768]: DEBUG nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Instance network_info: |[{"id": "51ac6b12-5631-4ce5-acc4-2f43dee834d0", "address": "fa:16:3e:82:95:dd", "network": {"id": "2fd0547b-2e4e-4625-9281-1c43c9b5ceee", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-186011702-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "9f4804c086b449608ef34ea22c018907", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "dd7d0d95-6848-4e69-ac21-75f8db82a3b5", "external-id": "nsx-vlan-transportzone-272", "segmentation_id": 272, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51ac6b12-56", "ovs_interfaceid": "51ac6b12-5631-4ce5-acc4-2f43dee834d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 930.040889] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance df570987-0144-4e3a-83db-a9538d3835f5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 930.044949] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:82:95:dd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'dd7d0d95-6848-4e69-ac21-75f8db82a3b5', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51ac6b12-5631-4ce5-acc4-2f43dee834d0', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 930.049276] env[61768]: DEBUG oslo.service.loopingcall [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.053344] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 930.053557] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 930.053697] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 930.055786] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 930.057184] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-25af9a57-75ee-4984-ae1c-30cff10cde24 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.077761] env[61768]: DEBUG oslo_vmware.api [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229218, 'name': PowerOffVM_Task, 'duration_secs': 0.206211} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.078078] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 930.078270] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 930.078739] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4e8f600a-0828-4820-b15e-12c7e83f1177 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.083578] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 930.083578] env[61768]: value = "task-1229219" [ 930.083578] env[61768]: _type = "Task" [ 930.083578] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.095207] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229219, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.170437] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 930.170731] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 930.170864] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleting the datastore file [datastore2] e0c89fa6-9d32-4343-89d7-e9129d252b40 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.171406] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee6421a1-e0d0-4111-b500-38cb42237041 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.180508] env[61768]: DEBUG oslo_vmware.api [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 930.180508] env[61768]: value = "task-1229221" [ 930.180508] env[61768]: _type = "Task" [ 930.180508] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.193674] env[61768]: DEBUG oslo_vmware.api [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229221, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.260646] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2d2fc51-7f44-482c-b6df-a6de01f1f4d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.269302] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118f8016-e3bc-4cec-99b0-bb8a670c479c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.299576] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e861cf-2efb-4752-9eac-8c7a2d73777b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.307553] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec38f1a8-2bca-42d0-9020-72f5b11e2452 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.323283] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.543944] env[61768]: DEBUG nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 930.574178] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 930.574995] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 930.574995] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 930.575264] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 930.575564] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 930.575835] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 930.576197] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 930.576473] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 930.576764] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 930.577061] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 930.577429] env[61768]: DEBUG nova.virt.hardware [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 930.578617] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f289b535-9e47-40f2-b526-1d05cdd7c8fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.591610] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdd8053-b344-4155-81fc-9623f5e8e900 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.598931] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229219, 'name': CreateVM_Task, 'duration_secs': 0.383762} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.599362] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 930.600113] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.600379] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.600697] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 930.600961] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b58e5616-1e27-4a56-9ff6-6dd25fe2c0b4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.614855] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 930.614855] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52256614-b90c-6604-e85f-7163d222f220" [ 930.614855] env[61768]: _type = "Task" [ 930.614855] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.622895] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52256614-b90c-6604-e85f-7163d222f220, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.691936] env[61768]: DEBUG oslo_vmware.api [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229221, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.16249} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.693284] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 930.693487] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 930.693671] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 930.693845] env[61768]: INFO nova.compute.manager [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Took 1.63 seconds to destroy the instance on the hypervisor. [ 930.694103] env[61768]: DEBUG oslo.service.loopingcall [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 930.695153] env[61768]: DEBUG nova.compute.manager [req-3d192fd3-e98e-4cac-8c78-295108fbc0aa req-997b8a00-ed2e-43ab-9169-384057d4ad49 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Received event network-vif-plugged-edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.695356] env[61768]: DEBUG oslo_concurrency.lockutils [req-3d192fd3-e98e-4cac-8c78-295108fbc0aa req-997b8a00-ed2e-43ab-9169-384057d4ad49 service nova] Acquiring lock "1158fbc3-f17b-44f7-847f-01bdf328a74a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.695563] env[61768]: DEBUG oslo_concurrency.lockutils [req-3d192fd3-e98e-4cac-8c78-295108fbc0aa req-997b8a00-ed2e-43ab-9169-384057d4ad49 service nova] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.695734] env[61768]: DEBUG oslo_concurrency.lockutils [req-3d192fd3-e98e-4cac-8c78-295108fbc0aa req-997b8a00-ed2e-43ab-9169-384057d4ad49 service nova] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.695903] env[61768]: DEBUG nova.compute.manager [req-3d192fd3-e98e-4cac-8c78-295108fbc0aa req-997b8a00-ed2e-43ab-9169-384057d4ad49 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] No waiting events found dispatching network-vif-plugged-edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 930.696200] env[61768]: WARNING nova.compute.manager [req-3d192fd3-e98e-4cac-8c78-295108fbc0aa req-997b8a00-ed2e-43ab-9169-384057d4ad49 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Received unexpected event network-vif-plugged-edd83e46-b8b7-4276-b563-16d34bb22e9b for instance with vm_state building and task_state spawning. [ 930.696493] env[61768]: DEBUG nova.compute.manager [-] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 930.696593] env[61768]: DEBUG nova.network.neutron [-] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 930.794586] env[61768]: DEBUG nova.network.neutron [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Successfully updated port: edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.826113] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.997437] env[61768]: DEBUG nova.compute.manager [req-73419063-b245-4e9c-9149-68ba70b3bf7a req-f751fbcf-4895-4de7-8b08-cac7ffb367fb service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Received event network-vif-deleted-5b0fd366-9b1d-4136-8996-73c527e03783 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.997682] env[61768]: INFO nova.compute.manager [req-73419063-b245-4e9c-9149-68ba70b3bf7a req-f751fbcf-4895-4de7-8b08-cac7ffb367fb service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Neutron deleted interface 5b0fd366-9b1d-4136-8996-73c527e03783; detaching it from the instance and deleting it from the info cache [ 930.998079] env[61768]: DEBUG nova.network.neutron [req-73419063-b245-4e9c-9149-68ba70b3bf7a req-f751fbcf-4895-4de7-8b08-cac7ffb367fb service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.128760] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52256614-b90c-6604-e85f-7163d222f220, 'name': SearchDatastore_Task, 'duration_secs': 0.010211} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.129174] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.129580] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 931.129993] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.130195] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.130399] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.131047] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-de98a2c2-db24-4d49-8e65-7eb32c3556da {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.141037] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.141037] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 931.141745] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca611e1d-9bdb-4c5b-befa-02acdfd6a004 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.147521] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 931.147521] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5278153f-519a-301d-961f-509a10c6f5dc" [ 931.147521] env[61768]: _type = "Task" [ 931.147521] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.157161] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5278153f-519a-301d-961f-509a10c6f5dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.300339] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.300565] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquired lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.300698] env[61768]: DEBUG nova.network.neutron [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 931.330740] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 931.330912] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.109s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.331320] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.661s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.332798] env[61768]: INFO nova.compute.claims [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 931.477259] env[61768]: DEBUG nova.network.neutron [-] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.500425] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0481267f-2bbc-4eb0-897b-0b3101839242 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.510614] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffba239d-6dd8-4b4a-a087-3f1ada077648 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.539704] env[61768]: DEBUG nova.compute.manager [req-73419063-b245-4e9c-9149-68ba70b3bf7a req-f751fbcf-4895-4de7-8b08-cac7ffb367fb service nova] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Detach interface failed, port_id=5b0fd366-9b1d-4136-8996-73c527e03783, reason: Instance e0c89fa6-9d32-4343-89d7-e9129d252b40 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 931.657608] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5278153f-519a-301d-961f-509a10c6f5dc, 'name': SearchDatastore_Task, 'duration_secs': 0.011189} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.658390] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b8b483a-0f83-4f5d-99db-ec583aff9c4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.663502] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 931.663502] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f0829c-2a1b-d932-1ead-4b771286bd7d" [ 931.663502] env[61768]: _type = "Task" [ 931.663502] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.670924] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f0829c-2a1b-d932-1ead-4b771286bd7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.834407] env[61768]: DEBUG nova.network.neutron [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 931.973217] env[61768]: DEBUG nova.network.neutron [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updating instance_info_cache with network_info: [{"id": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "address": "fa:16:3e:a7:9e:51", "network": {"id": "07321fe6-3bc8-41ee-9485-7873816a3b2e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-669074208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64f5133fd59844859b8e01d31f3320ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedd83e46-b8", "ovs_interfaceid": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.980657] env[61768]: INFO nova.compute.manager [-] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Took 1.28 seconds to deallocate network for instance. [ 932.174849] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f0829c-2a1b-d932-1ead-4b771286bd7d, 'name': SearchDatastore_Task, 'duration_secs': 0.01035} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.175182] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.175364] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] fb344f81-9295-44a1-9bac-dd530c157e98/fb344f81-9295-44a1-9bac-dd530c157e98.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 932.175636] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9fe5955e-6f8c-4b75-95fe-57f3fa809731 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.183435] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 932.183435] env[61768]: value = "task-1229222" [ 932.183435] env[61768]: _type = "Task" [ 932.183435] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.192951] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229222, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.478484] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Releasing lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.478892] env[61768]: DEBUG nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Instance network_info: |[{"id": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "address": "fa:16:3e:a7:9e:51", "network": {"id": "07321fe6-3bc8-41ee-9485-7873816a3b2e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-669074208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64f5133fd59844859b8e01d31f3320ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedd83e46-b8", "ovs_interfaceid": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 932.479648] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:9e:51', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea4fe416-47a6-4542-b59d-8c71ab4d6503', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'edd83e46-b8b7-4276-b563-16d34bb22e9b', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.487374] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Creating folder: Project (64f5133fd59844859b8e01d31f3320ab). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 932.488426] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.488755] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-877f6f66-318f-4d8d-a443-3728d880caec {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.509631] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Created folder: Project (64f5133fd59844859b8e01d31f3320ab) in parent group-v265360. [ 932.509951] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Creating folder: Instances. Parent ref: group-v265509. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 932.510206] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f33f154-df3a-4867-8bc7-106cbdf1f622 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.527423] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Created folder: Instances in parent group-v265509. [ 932.527771] env[61768]: DEBUG oslo.service.loopingcall [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.528241] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 932.528725] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a1e272bf-a825-4e07-94f8-b44cb8331765 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.556065] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.556065] env[61768]: value = "task-1229225" [ 932.556065] env[61768]: _type = "Task" [ 932.556065] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.561588] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d01b546-c9d3-4001-85a4-9ef851eca673 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.570193] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229225, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.573352] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0455062c-c2bb-438a-b30e-7ad5e985ed44 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.608027] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a0a0a7-02ae-4bc6-822b-ee2518072fda {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.616749] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7efde7b0-65b9-48e4-9839-bb8b18111952 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.631350] env[61768]: DEBUG nova.compute.provider_tree [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.695410] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229222, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461025} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.695623] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] fb344f81-9295-44a1-9bac-dd530c157e98/fb344f81-9295-44a1-9bac-dd530c157e98.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 932.695843] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 932.696223] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5923feaf-ce14-462a-b9a7-7d5aad9d5730 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.703364] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 932.703364] env[61768]: value = "task-1229226" [ 932.703364] env[61768]: _type = "Task" [ 932.703364] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.712666] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229226, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.722177] env[61768]: DEBUG nova.compute.manager [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Received event network-changed-edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.722410] env[61768]: DEBUG nova.compute.manager [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Refreshing instance network info cache due to event network-changed-edd83e46-b8b7-4276-b563-16d34bb22e9b. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.722674] env[61768]: DEBUG oslo_concurrency.lockutils [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] Acquiring lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.722767] env[61768]: DEBUG oslo_concurrency.lockutils [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] Acquired lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.722922] env[61768]: DEBUG nova.network.neutron [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Refreshing network info cache for port edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 932.756647] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.756968] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.066944] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229225, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.135304] env[61768]: DEBUG nova.scheduler.client.report [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.213558] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229226, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069013} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.213957] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 933.214732] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46da263d-f19a-4940-9ac7-16e346a3d3de {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.239084] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] fb344f81-9295-44a1-9bac-dd530c157e98/fb344f81-9295-44a1-9bac-dd530c157e98.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 933.239338] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0a84bc2f-feed-4aae-8ac8-b0718acd8385 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.260271] env[61768]: INFO nova.compute.manager [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Detaching volume b2dc03b4-7708-4dfc-ba53-f371792dd1ae [ 933.263986] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 933.263986] env[61768]: value = "task-1229227" [ 933.263986] env[61768]: _type = "Task" [ 933.263986] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.276131] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229227, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.296748] env[61768]: INFO nova.virt.block_device [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Attempting to driver detach volume b2dc03b4-7708-4dfc-ba53-f371792dd1ae from mountpoint /dev/sdb [ 933.297842] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 933.297842] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265482', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'name': 'volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '62cddff5-d499-4ef3-869d-3cdb0328640d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'serial': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 933.298406] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfab514-6cf3-4065-b334-2faa620280c9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.321693] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92dff8d7-045d-44dc-9b7e-71cd8aa44a95 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.329925] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab4418bc-0413-4b07-83ea-f46de36bf979 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.355573] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-539df855-6d28-46e4-b89f-6576316a001e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.374008] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] The volume has not been displaced from its original location: [datastore2] volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae/volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 933.379418] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfiguring VM instance instance-0000003e to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 933.379756] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcf79164-8e29-4c8c-817d-fb313f75e436 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.400914] env[61768]: DEBUG oslo_vmware.api [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 933.400914] env[61768]: value = "task-1229228" [ 933.400914] env[61768]: _type = "Task" [ 933.400914] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.409736] env[61768]: DEBUG oslo_vmware.api [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229228, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.511689] env[61768]: DEBUG nova.network.neutron [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updated VIF entry in instance network info cache for port edd83e46-b8b7-4276-b563-16d34bb22e9b. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 933.512206] env[61768]: DEBUG nova.network.neutron [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updating instance_info_cache with network_info: [{"id": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "address": "fa:16:3e:a7:9e:51", "network": {"id": "07321fe6-3bc8-41ee-9485-7873816a3b2e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-669074208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64f5133fd59844859b8e01d31f3320ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedd83e46-b8", "ovs_interfaceid": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.567463] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229225, 'name': CreateVM_Task, 'duration_secs': 0.52027} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.567634] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 933.568450] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.568645] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.568965] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.569252] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67e1e964-c9cd-4994-abb1-bc55fba8991f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.574621] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 933.574621] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52348ee9-d34e-a7d7-640d-53e5d0877e9e" [ 933.574621] env[61768]: _type = "Task" [ 933.574621] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.583811] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52348ee9-d34e-a7d7-640d-53e5d0877e9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.640891] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.641055] env[61768]: DEBUG nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 933.643644] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.325s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.643820] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.645834] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.212s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.647209] env[61768]: INFO nova.compute.claims [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.669014] env[61768]: INFO nova.scheduler.client.report [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleted allocations for instance 5cc705a5-1423-43e1-8356-9abd67a8da3a [ 933.774720] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229227, 'name': ReconfigVM_Task, 'duration_secs': 0.462135} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.775134] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Reconfigured VM instance instance-0000005d to attach disk [datastore2] fb344f81-9295-44a1-9bac-dd530c157e98/fb344f81-9295-44a1-9bac-dd530c157e98.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 933.775789] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89f58bc5-0acf-4908-8bcd-adf1a5f83fe8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.782618] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 933.782618] env[61768]: value = "task-1229229" [ 933.782618] env[61768]: _type = "Task" [ 933.782618] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.790958] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229229, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.912656] env[61768]: DEBUG oslo_vmware.api [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229228, 'name': ReconfigVM_Task, 'duration_secs': 0.254014} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.912967] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Reconfigured VM instance instance-0000003e to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.917531] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-558cc58b-a27c-46ec-8302-ab537a67ba74 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.933793] env[61768]: DEBUG oslo_vmware.api [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 933.933793] env[61768]: value = "task-1229230" [ 933.933793] env[61768]: _type = "Task" [ 933.933793] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.942648] env[61768]: DEBUG oslo_vmware.api [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229230, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.015246] env[61768]: DEBUG oslo_concurrency.lockutils [req-08e96265-2ecc-4c43-8932-00da86ae5de7 req-4e3ea58d-2c43-4407-8590-083173eddae8 service nova] Releasing lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.044172] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.044662] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.044996] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.045367] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.045599] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.048219] env[61768]: INFO nova.compute.manager [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Terminating instance [ 934.050234] env[61768]: DEBUG nova.compute.manager [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 934.050458] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 934.051394] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8d5949-b51f-443a-a387-6eedb99d4d31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.059706] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 934.060063] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdf15f4d-a142-4239-9320-f7620a59f69e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.067153] env[61768]: DEBUG oslo_vmware.api [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 934.067153] env[61768]: value = "task-1229231" [ 934.067153] env[61768]: _type = "Task" [ 934.067153] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.075515] env[61768]: DEBUG oslo_vmware.api [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229231, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.084134] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52348ee9-d34e-a7d7-640d-53e5d0877e9e, 'name': SearchDatastore_Task, 'duration_secs': 0.012235} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.085169] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.085169] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 934.085169] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.085169] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.085468] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 934.085538] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef0edb01-f3b7-4ae8-8d34-937f711c23a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.095541] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 934.095785] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 934.096504] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e4af33f-fabf-41db-90de-7930ca442e09 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.103190] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 934.103190] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d74c2-1574-7fed-5d45-5a1248273823" [ 934.103190] env[61768]: _type = "Task" [ 934.103190] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.110231] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d74c2-1574-7fed-5d45-5a1248273823, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.151881] env[61768]: DEBUG nova.compute.utils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 934.155473] env[61768]: DEBUG nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 934.155696] env[61768]: DEBUG nova.network.neutron [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 934.177216] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b8dff17e-1758-4f5f-9991-38da08f76b7e tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "5cc705a5-1423-43e1-8356-9abd67a8da3a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.428s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.212845] env[61768]: DEBUG nova.policy [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1b1f8eecca543e094a73fef9b6707f9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '98ece3463c0f42d290a78f515efeb52f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 934.294467] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229229, 'name': Rename_Task, 'duration_secs': 0.139465} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.294856] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 934.295162] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-df7b1bda-34be-4b00-a6c7-11048f564fab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.303185] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 934.303185] env[61768]: value = "task-1229232" [ 934.303185] env[61768]: _type = "Task" [ 934.303185] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.313141] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229232, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.450856] env[61768]: DEBUG oslo_vmware.api [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229230, 'name': ReconfigVM_Task, 'duration_secs': 0.135751} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.451276] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265482', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'name': 'volume-b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '62cddff5-d499-4ef3-869d-3cdb0328640d', 'attached_at': '', 'detached_at': '', 'volume_id': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae', 'serial': 'b2dc03b4-7708-4dfc-ba53-f371792dd1ae'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 934.465894] env[61768]: DEBUG nova.network.neutron [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Successfully created port: 94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 934.578351] env[61768]: DEBUG oslo_vmware.api [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229231, 'name': PowerOffVM_Task, 'duration_secs': 0.245618} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.578630] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 934.578805] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 934.579078] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e0f4bd9-5b88-44bf-94be-1c2c5b85c577 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.612990] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529d74c2-1574-7fed-5d45-5a1248273823, 'name': SearchDatastore_Task, 'duration_secs': 0.014196} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.613811] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76ea8697-68f7-4eca-b4b2-0e103c9e79cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.621026] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 934.621026] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52969730-6746-0185-a2cc-4a78de8ad9ad" [ 934.621026] env[61768]: _type = "Task" [ 934.621026] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.631360] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52969730-6746-0185-a2cc-4a78de8ad9ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.646171] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 934.646514] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 934.646635] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Deleting the datastore file [datastore1] 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.646857] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aeded01a-56f8-46f6-b990-a163e0f9ab0b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.654159] env[61768]: DEBUG oslo_vmware.api [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 934.654159] env[61768]: value = "task-1229234" [ 934.654159] env[61768]: _type = "Task" [ 934.654159] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.657580] env[61768]: DEBUG nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 934.670647] env[61768]: DEBUG oslo_vmware.api [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229234, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.816485] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229232, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.864094] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc072b2-cdba-4719-81fa-0a4f9de2d6e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.877388] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63628807-f7b8-4928-917f-c9c41a9921c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.912263] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99955f30-b261-4f43-8586-85138e4e6358 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.920077] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9816848-ca5d-4a1f-a772-8743d94320c7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.933472] env[61768]: DEBUG nova.compute.provider_tree [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.976235] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "ecd08348-b22f-49c5-b13f-b187506e38d6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.976504] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.976719] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.976911] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.977109] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.979290] env[61768]: INFO nova.compute.manager [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Terminating instance [ 934.981169] env[61768]: DEBUG nova.compute.manager [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 934.981384] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 934.982226] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaf88639-4de0-468d-8242-35acc987d894 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.991127] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 934.991127] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7f6088c5-bcb7-464c-957b-2a8e0a072538 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.999018] env[61768]: DEBUG oslo_vmware.api [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 934.999018] env[61768]: value = "task-1229235" [ 934.999018] env[61768]: _type = "Task" [ 934.999018] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.008590] env[61768]: DEBUG oslo_vmware.api [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229235, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.020918] env[61768]: DEBUG nova.objects.instance [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'flavor' on Instance uuid 62cddff5-d499-4ef3-869d-3cdb0328640d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.134236] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52969730-6746-0185-a2cc-4a78de8ad9ad, 'name': SearchDatastore_Task, 'duration_secs': 0.018628} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.134545] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.134833] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 1158fbc3-f17b-44f7-847f-01bdf328a74a/1158fbc3-f17b-44f7-847f-01bdf328a74a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 935.135123] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-145d4fd0-0aef-4217-a06c-29eb388c478e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.143989] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 935.143989] env[61768]: value = "task-1229236" [ 935.143989] env[61768]: _type = "Task" [ 935.143989] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.154417] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229236, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.163853] env[61768]: DEBUG oslo_vmware.api [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229234, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239241} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.164176] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.164388] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 935.164595] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 935.164827] env[61768]: INFO nova.compute.manager [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Took 1.11 seconds to destroy the instance on the hypervisor. [ 935.165150] env[61768]: DEBUG oslo.service.loopingcall [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 935.165335] env[61768]: DEBUG nova.compute.manager [-] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 935.165437] env[61768]: DEBUG nova.network.neutron [-] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 935.314704] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229232, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.437749] env[61768]: DEBUG nova.scheduler.client.report [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.515875] env[61768]: DEBUG oslo_vmware.api [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229235, 'name': PowerOffVM_Task, 'duration_secs': 0.279254} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.516259] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 935.516427] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 935.516749] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-42e82a97-21ad-4891-986d-0a7bb3e050cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.602291] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 935.605297] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 935.605297] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleting the datastore file [datastore2] ecd08348-b22f-49c5-b13f-b187506e38d6 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 935.605297] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9b6dce4e-023e-47ff-801b-1c8520ddeca2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.613680] env[61768]: DEBUG oslo_vmware.api [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for the task: (returnval){ [ 935.613680] env[61768]: value = "task-1229238" [ 935.613680] env[61768]: _type = "Task" [ 935.613680] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.623446] env[61768]: DEBUG oslo_vmware.api [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229238, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.656685] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229236, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.672530] env[61768]: DEBUG nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 935.701222] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 935.701519] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 935.701687] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 935.701881] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 935.702048] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 935.702215] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 935.702476] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 935.702637] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 935.703049] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 935.703243] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 935.703425] env[61768]: DEBUG nova.virt.hardware [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 935.704307] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036202da-61f1-45c7-b716-0794c58e2630 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.717125] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f02e1a-f670-4caf-bbb8-a7cebd34cd7a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.815777] env[61768]: DEBUG oslo_vmware.api [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229232, 'name': PowerOnVM_Task, 'duration_secs': 1.190702} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.817878] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 935.817878] env[61768]: INFO nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Took 9.67 seconds to spawn the instance on the hypervisor. [ 935.817878] env[61768]: DEBUG nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.817878] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38eab0c-9570-4270-b1fd-7e8db205dca3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.884398] env[61768]: DEBUG nova.compute.manager [req-77d8bcc8-9460-42ff-a70c-2c2165197e50 req-37003770-fd1a-4bc6-a4f2-55cfb1f1d02a service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Received event network-vif-deleted-51b2319f-ceb1-41db-8019-1d918e70080d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 935.884678] env[61768]: INFO nova.compute.manager [req-77d8bcc8-9460-42ff-a70c-2c2165197e50 req-37003770-fd1a-4bc6-a4f2-55cfb1f1d02a service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Neutron deleted interface 51b2319f-ceb1-41db-8019-1d918e70080d; detaching it from the instance and deleting it from the info cache [ 935.884907] env[61768]: DEBUG nova.network.neutron [req-77d8bcc8-9460-42ff-a70c-2c2165197e50 req-37003770-fd1a-4bc6-a4f2-55cfb1f1d02a service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.943087] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.943595] env[61768]: DEBUG nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.946961] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.459s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.947415] env[61768]: DEBUG nova.objects.instance [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'resources' on Instance uuid e0c89fa6-9d32-4343-89d7-e9129d252b40 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.031914] env[61768]: DEBUG oslo_concurrency.lockutils [None req-59a52abe-c2ab-487d-8bb2-c5c8b63cb2d5 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.274s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.124207] env[61768]: DEBUG oslo_vmware.api [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Task: {'id': task-1229238, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196143} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.124531] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 936.124727] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 936.124961] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 936.125105] env[61768]: INFO nova.compute.manager [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 936.125355] env[61768]: DEBUG oslo.service.loopingcall [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 936.125555] env[61768]: DEBUG nova.compute.manager [-] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 936.125647] env[61768]: DEBUG nova.network.neutron [-] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 936.154943] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229236, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540456} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.155239] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 1158fbc3-f17b-44f7-847f-01bdf328a74a/1158fbc3-f17b-44f7-847f-01bdf328a74a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 936.155572] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.155715] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-47ec7a44-1ec3-4945-99e1-24907a273ba5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.164128] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 936.164128] env[61768]: value = "task-1229239" [ 936.164128] env[61768]: _type = "Task" [ 936.164128] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.173956] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229239, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.215850] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.216341] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.003s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.216679] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.216935] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.217162] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.219756] env[61768]: INFO nova.compute.manager [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Terminating instance [ 936.221815] env[61768]: DEBUG nova.compute.manager [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 936.222161] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 936.222898] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea74dff-701c-4c88-aedb-51c02e7f7da9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.236075] env[61768]: DEBUG nova.network.neutron [-] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.240417] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 936.240417] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-06d8abe0-2111-4f84-a419-dba89c0f1869 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.246612] env[61768]: DEBUG oslo_vmware.api [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 936.246612] env[61768]: value = "task-1229240" [ 936.246612] env[61768]: _type = "Task" [ 936.246612] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.257799] env[61768]: DEBUG oslo_vmware.api [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229240, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.275346] env[61768]: DEBUG nova.compute.manager [req-75fd4a2e-157f-44d2-b24d-7f30a4049d9e req-16e7f6c0-facc-41d2-82a6-b2db9d968d5b service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Received event network-vif-plugged-94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 936.275346] env[61768]: DEBUG oslo_concurrency.lockutils [req-75fd4a2e-157f-44d2-b24d-7f30a4049d9e req-16e7f6c0-facc-41d2-82a6-b2db9d968d5b service nova] Acquiring lock "df570987-0144-4e3a-83db-a9538d3835f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.275346] env[61768]: DEBUG oslo_concurrency.lockutils [req-75fd4a2e-157f-44d2-b24d-7f30a4049d9e req-16e7f6c0-facc-41d2-82a6-b2db9d968d5b service nova] Lock "df570987-0144-4e3a-83db-a9538d3835f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.275635] env[61768]: DEBUG oslo_concurrency.lockutils [req-75fd4a2e-157f-44d2-b24d-7f30a4049d9e req-16e7f6c0-facc-41d2-82a6-b2db9d968d5b service nova] Lock "df570987-0144-4e3a-83db-a9538d3835f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.275687] env[61768]: DEBUG nova.compute.manager [req-75fd4a2e-157f-44d2-b24d-7f30a4049d9e req-16e7f6c0-facc-41d2-82a6-b2db9d968d5b service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] No waiting events found dispatching network-vif-plugged-94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 936.275859] env[61768]: WARNING nova.compute.manager [req-75fd4a2e-157f-44d2-b24d-7f30a4049d9e req-16e7f6c0-facc-41d2-82a6-b2db9d968d5b service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Received unexpected event network-vif-plugged-94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 for instance with vm_state building and task_state spawning. [ 936.337280] env[61768]: INFO nova.compute.manager [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Took 19.11 seconds to build instance. [ 936.387803] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1159575e-aa47-4f66-8a52-24c314973c69 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.399851] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce1f81a0-c4f0-4a84-b483-1015882ea8b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.413438] env[61768]: DEBUG nova.network.neutron [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Successfully updated port: 94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 936.439826] env[61768]: DEBUG nova.compute.manager [req-77d8bcc8-9460-42ff-a70c-2c2165197e50 req-37003770-fd1a-4bc6-a4f2-55cfb1f1d02a service nova] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Detach interface failed, port_id=51b2319f-ceb1-41db-8019-1d918e70080d, reason: Instance 7697ecda-ef30-447d-a92a-3f5cb3cc9118 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 936.453886] env[61768]: DEBUG nova.compute.utils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.457895] env[61768]: DEBUG nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 936.458096] env[61768]: DEBUG nova.network.neutron [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 936.502538] env[61768]: DEBUG nova.policy [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd1aa7c174648888b58f6a6df6f82a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfce0cf4a5d14042a929dac65aeb2ae3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.642386] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6928d560-deaa-403b-95f5-4baa141f33d6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.650462] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a65414b-6395-451a-96d2-6cc40fc3a1cd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.683385] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c88a23c-fc4a-4a1c-b76a-0fae23014721 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.693989] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6820759-25a2-4915-bb69-9ac64cd3ca79 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.698690] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229239, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07594} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.698992] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 936.700278] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1da5d3c-aab6-4dd9-b68b-5993fc53ef5c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.714227] env[61768]: DEBUG nova.compute.provider_tree [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.738703] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] 1158fbc3-f17b-44f7-847f-01bdf328a74a/1158fbc3-f17b-44f7-847f-01bdf328a74a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 936.740059] env[61768]: INFO nova.compute.manager [-] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Took 1.57 seconds to deallocate network for instance. [ 936.740322] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-331efb5f-db4a-429c-ba97-ffb50bce1b0a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.755229] env[61768]: DEBUG nova.network.neutron [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Successfully created port: 9681c697-a94e-4494-a310-1468fe766740 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.773633] env[61768]: DEBUG oslo_vmware.api [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229240, 'name': PowerOffVM_Task, 'duration_secs': 0.49079} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.775922] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 936.775922] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 936.775922] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 936.775922] env[61768]: value = "task-1229241" [ 936.775922] env[61768]: _type = "Task" [ 936.775922] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.775922] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df709c3e-d24b-4525-abe6-b4040e4cb32e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.785916] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229241, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.839603] env[61768]: DEBUG oslo_concurrency.lockutils [None req-edfd7ad7-b85e-4c7f-b5e6-4a27f88bea4e tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "fb344f81-9295-44a1-9bac-dd530c157e98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.621s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.857087] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 936.857087] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 936.857087] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleting the datastore file [datastore2] 62cddff5-d499-4ef3-869d-3cdb0328640d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 936.857087] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d03a22f3-3c71-45cc-981e-1981d91a003e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.864063] env[61768]: DEBUG oslo_vmware.api [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 936.864063] env[61768]: value = "task-1229243" [ 936.864063] env[61768]: _type = "Task" [ 936.864063] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.874158] env[61768]: DEBUG oslo_vmware.api [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229243, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.883514] env[61768]: DEBUG nova.network.neutron [-] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.916857] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "refresh_cache-df570987-0144-4e3a-83db-a9538d3835f5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.916857] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquired lock "refresh_cache-df570987-0144-4e3a-83db-a9538d3835f5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.916857] env[61768]: DEBUG nova.network.neutron [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 936.958606] env[61768]: DEBUG nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 937.218129] env[61768]: DEBUG nova.scheduler.client.report [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.264225] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.290633] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229241, 'name': ReconfigVM_Task, 'duration_secs': 0.317527} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.291059] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Reconfigured VM instance instance-0000005e to attach disk [datastore2] 1158fbc3-f17b-44f7-847f-01bdf328a74a/1158fbc3-f17b-44f7-847f-01bdf328a74a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 937.292044] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d6af62f3-14d0-4152-a935-74509d33f1ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.300737] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 937.300737] env[61768]: value = "task-1229244" [ 937.300737] env[61768]: _type = "Task" [ 937.300737] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.312861] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229244, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.377370] env[61768]: DEBUG oslo_vmware.api [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229243, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143606} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.377652] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 937.377849] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 937.378202] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 937.378266] env[61768]: INFO nova.compute.manager [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 937.378517] env[61768]: DEBUG oslo.service.loopingcall [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.378712] env[61768]: DEBUG nova.compute.manager [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 937.378814] env[61768]: DEBUG nova.network.neutron [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 937.386198] env[61768]: INFO nova.compute.manager [-] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Took 1.26 seconds to deallocate network for instance. [ 937.467784] env[61768]: DEBUG nova.network.neutron [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 937.725085] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.778s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.729130] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.464s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.732930] env[61768]: DEBUG nova.objects.instance [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'resources' on Instance uuid 7697ecda-ef30-447d-a92a-3f5cb3cc9118 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 937.752413] env[61768]: INFO nova.scheduler.client.report [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted allocations for instance e0c89fa6-9d32-4343-89d7-e9129d252b40 [ 937.776536] env[61768]: DEBUG nova.network.neutron [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Updating instance_info_cache with network_info: [{"id": "94ba3a4c-7b23-48e1-a23d-bef8d5eaa582", "address": "fa:16:3e:4b:4d:05", "network": {"id": "c72d6dfa-5af7-4d7a-8ac9-721c7418c93d", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1886422157-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98ece3463c0f42d290a78f515efeb52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ba3a4c-7b", "ovs_interfaceid": "94ba3a4c-7b23-48e1-a23d-bef8d5eaa582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 937.812147] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229244, 'name': Rename_Task, 'duration_secs': 0.149761} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.812473] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 937.812731] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9d8d4f9f-4fd8-4760-aadb-cf8986ffe031 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.821388] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 937.821388] env[61768]: value = "task-1229245" [ 937.821388] env[61768]: _type = "Task" [ 937.821388] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.830969] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229245, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.892845] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.949102] env[61768]: DEBUG nova.compute.manager [req-fca41cea-8c54-44f5-be7e-0358610f69e7 req-31c5238e-a6a4-4ae2-9d62-35377f28602d service nova] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Received event network-vif-deleted-b4ad17b4-d077-472c-be99-bd42736950e4 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 937.978423] env[61768]: DEBUG nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 938.002635] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 938.003175] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 938.003558] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 938.003920] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 938.004277] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 938.004641] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 938.005036] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 938.005378] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 938.005717] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 938.007025] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 938.007025] env[61768]: DEBUG nova.virt.hardware [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 938.007964] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57fa4a7b-1f2a-4ee4-b550-8d831c638d63 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.017824] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0584145-c617-475e-b08c-856cb8c1d487 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.260616] env[61768]: DEBUG oslo_concurrency.lockutils [None req-be167f56-4712-4baf-bcaf-97f07417e314 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0c89fa6-9d32-4343-89d7-e9129d252b40" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.198s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.279466] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Releasing lock "refresh_cache-df570987-0144-4e3a-83db-a9538d3835f5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.280320] env[61768]: DEBUG nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Instance network_info: |[{"id": "94ba3a4c-7b23-48e1-a23d-bef8d5eaa582", "address": "fa:16:3e:4b:4d:05", "network": {"id": "c72d6dfa-5af7-4d7a-8ac9-721c7418c93d", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1886422157-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98ece3463c0f42d290a78f515efeb52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ba3a4c-7b", "ovs_interfaceid": "94ba3a4c-7b23-48e1-a23d-bef8d5eaa582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 938.281052] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:4d:05', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2ed91b7b-b4ec-486d-ab34-af0afb7ec691', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '94ba3a4c-7b23-48e1-a23d-bef8d5eaa582', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 938.288509] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Creating folder: Project (98ece3463c0f42d290a78f515efeb52f). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 938.289571] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e71f8c0e-4ace-4857-9878-a38c1c988c08 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.308311] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Created folder: Project (98ece3463c0f42d290a78f515efeb52f) in parent group-v265360. [ 938.308460] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Creating folder: Instances. Parent ref: group-v265512. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 938.308678] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb3bdf2b-8f56-4dcf-8cdc-49703ee64339 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.314823] env[61768]: DEBUG nova.compute.manager [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Received event network-changed-94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.315710] env[61768]: DEBUG nova.compute.manager [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Refreshing instance network info cache due to event network-changed-94ba3a4c-7b23-48e1-a23d-bef8d5eaa582. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 938.316187] env[61768]: DEBUG oslo_concurrency.lockutils [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] Acquiring lock "refresh_cache-df570987-0144-4e3a-83db-a9538d3835f5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.316414] env[61768]: DEBUG oslo_concurrency.lockutils [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] Acquired lock "refresh_cache-df570987-0144-4e3a-83db-a9538d3835f5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.317094] env[61768]: DEBUG nova.network.neutron [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Refreshing network info cache for port 94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 938.337918] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Created folder: Instances in parent group-v265512. [ 938.338595] env[61768]: DEBUG oslo.service.loopingcall [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 938.344636] env[61768]: DEBUG nova.compute.manager [req-03c08335-5fe5-44ad-b77d-d33950e454e1 req-e01ffa20-afcc-431e-9318-7453253c69ce service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Received event network-vif-plugged-9681c697-a94e-4494-a310-1468fe766740 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.345437] env[61768]: DEBUG oslo_concurrency.lockutils [req-03c08335-5fe5-44ad-b77d-d33950e454e1 req-e01ffa20-afcc-431e-9318-7453253c69ce service nova] Acquiring lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.345437] env[61768]: DEBUG oslo_concurrency.lockutils [req-03c08335-5fe5-44ad-b77d-d33950e454e1 req-e01ffa20-afcc-431e-9318-7453253c69ce service nova] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.345437] env[61768]: DEBUG oslo_concurrency.lockutils [req-03c08335-5fe5-44ad-b77d-d33950e454e1 req-e01ffa20-afcc-431e-9318-7453253c69ce service nova] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.345917] env[61768]: DEBUG nova.compute.manager [req-03c08335-5fe5-44ad-b77d-d33950e454e1 req-e01ffa20-afcc-431e-9318-7453253c69ce service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] No waiting events found dispatching network-vif-plugged-9681c697-a94e-4494-a310-1468fe766740 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.345917] env[61768]: WARNING nova.compute.manager [req-03c08335-5fe5-44ad-b77d-d33950e454e1 req-e01ffa20-afcc-431e-9318-7453253c69ce service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Received unexpected event network-vif-plugged-9681c697-a94e-4494-a310-1468fe766740 for instance with vm_state building and task_state spawning. [ 938.346512] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 938.347102] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229245, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.350831] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f3b588c-71c1-4b44-a218-a40e3fe980f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.372814] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "fb344f81-9295-44a1-9bac-dd530c157e98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.373975] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "fb344f81-9295-44a1-9bac-dd530c157e98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.373975] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "fb344f81-9295-44a1-9bac-dd530c157e98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.373975] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "fb344f81-9295-44a1-9bac-dd530c157e98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.373975] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "fb344f81-9295-44a1-9bac-dd530c157e98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.376328] env[61768]: INFO nova.compute.manager [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Terminating instance [ 938.378508] env[61768]: DEBUG nova.compute.manager [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.378842] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.380143] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8c166d-d47a-4976-a4ba-a7b62bfcdf5c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.387543] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 938.387543] env[61768]: value = "task-1229248" [ 938.387543] env[61768]: _type = "Task" [ 938.387543] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.400242] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.401395] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bc81a961-782e-4a55-b253-66a45c1722f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.408422] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229248, 'name': CreateVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.415043] env[61768]: DEBUG oslo_vmware.api [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 938.415043] env[61768]: value = "task-1229249" [ 938.415043] env[61768]: _type = "Task" [ 938.415043] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.426800] env[61768]: DEBUG oslo_vmware.api [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229249, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.512617] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca77206c-917b-4792-aa92-b28ce993b10f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.521090] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ee97bb-24ae-4e95-95a7-098514f6e8a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.554093] env[61768]: DEBUG nova.network.neutron [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.556804] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00038603-d8ad-4517-b5a5-4b60a37ba001 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.566502] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06935e15-d768-465a-9abb-db8c50d15014 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.582991] env[61768]: DEBUG nova.compute.provider_tree [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.834905] env[61768]: DEBUG oslo_vmware.api [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229245, 'name': PowerOnVM_Task, 'duration_secs': 0.593202} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.835293] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 938.835802] env[61768]: INFO nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Took 8.29 seconds to spawn the instance on the hypervisor. [ 938.836182] env[61768]: DEBUG nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 938.836960] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd133efa-8e8f-4e29-be0f-aa7caba4de1c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.901162] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229248, 'name': CreateVM_Task} progress is 25%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.928017] env[61768]: DEBUG oslo_vmware.api [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229249, 'name': PowerOffVM_Task, 'duration_secs': 0.21588} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.928819] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 938.928819] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 938.928939] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f85f81e0-88ce-4255-a703-d1daf73a6460 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.933097] env[61768]: DEBUG nova.network.neutron [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Successfully updated port: 9681c697-a94e-4494-a310-1468fe766740 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.955935] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "739faae0-5811-4f88-b56b-1350c0f7b8be" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.956331] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.956500] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "739faae0-5811-4f88-b56b-1350c0f7b8be-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.956761] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.956950] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.958950] env[61768]: INFO nova.compute.manager [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Terminating instance [ 938.960761] env[61768]: DEBUG nova.compute.manager [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 938.960950] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 938.962024] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f4e8e6-575f-4841-accf-5d14813177a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.972600] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 938.972875] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-00a24277-23cb-4a19-9729-7994cd0c116c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.992869] env[61768]: DEBUG oslo_vmware.api [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 938.992869] env[61768]: value = "task-1229251" [ 938.992869] env[61768]: _type = "Task" [ 938.992869] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.003796] env[61768]: DEBUG oslo_vmware.api [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.046109] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.046611] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.046782] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleting the datastore file [datastore2] fb344f81-9295-44a1-9bac-dd530c157e98 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.052538] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-51c17ee9-06ee-4abc-b9af-507a45f13aeb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.058029] env[61768]: INFO nova.compute.manager [-] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Took 1.68 seconds to deallocate network for instance. [ 939.071519] env[61768]: DEBUG oslo_vmware.api [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for the task: (returnval){ [ 939.071519] env[61768]: value = "task-1229252" [ 939.071519] env[61768]: _type = "Task" [ 939.071519] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.084026] env[61768]: DEBUG oslo_vmware.api [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229252, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.087074] env[61768]: DEBUG nova.scheduler.client.report [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.118318] env[61768]: DEBUG nova.network.neutron [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Updated VIF entry in instance network info cache for port 94ba3a4c-7b23-48e1-a23d-bef8d5eaa582. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 939.118837] env[61768]: DEBUG nova.network.neutron [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Updating instance_info_cache with network_info: [{"id": "94ba3a4c-7b23-48e1-a23d-bef8d5eaa582", "address": "fa:16:3e:4b:4d:05", "network": {"id": "c72d6dfa-5af7-4d7a-8ac9-721c7418c93d", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1886422157-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "98ece3463c0f42d290a78f515efeb52f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2ed91b7b-b4ec-486d-ab34-af0afb7ec691", "external-id": "nsx-vlan-transportzone-75", "segmentation_id": 75, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap94ba3a4c-7b", "ovs_interfaceid": "94ba3a4c-7b23-48e1-a23d-bef8d5eaa582", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.354978] env[61768]: INFO nova.compute.manager [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Took 18.01 seconds to build instance. [ 939.399574] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229248, 'name': CreateVM_Task, 'duration_secs': 0.683306} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.399920] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 939.400501] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.400667] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.400992] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 939.401326] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86bdb707-bb06-4523-9301-718999acb91e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.406985] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 939.406985] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52dfcda4-7ecd-12da-9741-170bde8dab2b" [ 939.406985] env[61768]: _type = "Task" [ 939.406985] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.414877] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52dfcda4-7ecd-12da-9741-170bde8dab2b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.435977] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.435977] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.435977] env[61768]: DEBUG nova.network.neutron [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.503160] env[61768]: DEBUG oslo_vmware.api [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229251, 'name': PowerOffVM_Task, 'duration_secs': 0.208418} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.503462] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 939.503635] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 939.503894] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-948e8e51-3177-4433-899e-a9c4eaf8301b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.571533] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.576864] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 939.577100] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 939.577304] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleting the datastore file [datastore1] 739faae0-5811-4f88-b56b-1350c0f7b8be {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 939.577966] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67aa2fed-a516-4e04-8259-87a7b08fc7d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.583272] env[61768]: DEBUG oslo_vmware.api [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Task: {'id': task-1229252, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.142683} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.584489] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 939.584681] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 939.584862] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 939.585060] env[61768]: INFO nova.compute.manager [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Took 1.21 seconds to destroy the instance on the hypervisor. [ 939.585343] env[61768]: DEBUG oslo.service.loopingcall [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 939.585602] env[61768]: DEBUG oslo_vmware.api [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 939.585602] env[61768]: value = "task-1229254" [ 939.585602] env[61768]: _type = "Task" [ 939.585602] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.585792] env[61768]: DEBUG nova.compute.manager [-] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 939.585888] env[61768]: DEBUG nova.network.neutron [-] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 939.592822] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.865s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.597644] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.705s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.597882] env[61768]: DEBUG nova.objects.instance [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lazy-loading 'resources' on Instance uuid ecd08348-b22f-49c5-b13f-b187506e38d6 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.599035] env[61768]: DEBUG oslo_vmware.api [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229254, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.617073] env[61768]: INFO nova.scheduler.client.report [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Deleted allocations for instance 7697ecda-ef30-447d-a92a-3f5cb3cc9118 [ 939.621940] env[61768]: DEBUG oslo_concurrency.lockutils [req-28117318-3811-4e39-9dc7-b0dffc8d3155 req-df87ffd7-3bea-4346-8d39-f047882249ae service nova] Releasing lock "refresh_cache-df570987-0144-4e3a-83db-a9538d3835f5" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.858203] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e430e088-bc45-4076-b0c9-0ef942951ce8 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.516s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.918221] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52dfcda4-7ecd-12da-9741-170bde8dab2b, 'name': SearchDatastore_Task, 'duration_secs': 0.009602} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.918552] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.918795] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 939.919067] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.919238] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.919432] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.919693] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5dfb0f39-8e97-4410-bc47-d95ddda34cc0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.928384] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.928572] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 939.929298] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1b5040b-6c83-4ab1-b677-d0e355208b52 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.934666] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 939.934666] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a50750-055d-f77d-04bf-afd182e734fa" [ 939.934666] env[61768]: _type = "Task" [ 939.934666] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.947268] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a50750-055d-f77d-04bf-afd182e734fa, 'name': SearchDatastore_Task, 'duration_secs': 0.008384} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.948161] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b240d08-5634-457e-a6aa-97a641a04f63 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.954772] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 939.954772] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5233fe61-a03a-a261-d423-f4f455c9cd4c" [ 939.954772] env[61768]: _type = "Task" [ 939.954772] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.965330] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5233fe61-a03a-a261-d423-f4f455c9cd4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.970799] env[61768]: DEBUG nova.network.neutron [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 939.976191] env[61768]: DEBUG nova.compute.manager [req-e3a68b1f-cb84-4ff2-9fd0-6b394720e054 req-a01694f9-4ea0-4382-9216-6dedf605526a service nova] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Received event network-vif-deleted-44b2acbd-7a9f-416b-b3b4-e877de5bfa85 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.976443] env[61768]: DEBUG nova.compute.manager [req-e3a68b1f-cb84-4ff2-9fd0-6b394720e054 req-a01694f9-4ea0-4382-9216-6dedf605526a service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Received event network-vif-deleted-51ac6b12-5631-4ce5-acc4-2f43dee834d0 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 939.976629] env[61768]: INFO nova.compute.manager [req-e3a68b1f-cb84-4ff2-9fd0-6b394720e054 req-a01694f9-4ea0-4382-9216-6dedf605526a service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Neutron deleted interface 51ac6b12-5631-4ce5-acc4-2f43dee834d0; detaching it from the instance and deleting it from the info cache [ 939.976804] env[61768]: DEBUG nova.network.neutron [req-e3a68b1f-cb84-4ff2-9fd0-6b394720e054 req-a01694f9-4ea0-4382-9216-6dedf605526a service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.102676] env[61768]: DEBUG oslo_vmware.api [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229254, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145339} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.103331] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 940.103538] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 940.103849] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 940.103911] env[61768]: INFO nova.compute.manager [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Took 1.14 seconds to destroy the instance on the hypervisor. [ 940.104162] env[61768]: DEBUG oslo.service.loopingcall [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.104366] env[61768]: DEBUG nova.compute.manager [-] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 940.104461] env[61768]: DEBUG nova.network.neutron [-] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 940.111868] env[61768]: DEBUG nova.network.neutron [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Updating instance_info_cache with network_info: [{"id": "9681c697-a94e-4494-a310-1468fe766740", "address": "fa:16:3e:c5:1d:3f", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9681c697-a9", "ovs_interfaceid": "9681c697-a94e-4494-a310-1468fe766740", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.125925] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f745c2a7-b28d-49d9-bf37-b172d7c240b4 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "7697ecda-ef30-447d-a92a-3f5cb3cc9118" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.081s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.273511] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d6d826-164f-4cd6-b605-ec688bf89f29 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.281886] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e8532b-89d7-4dcd-bf8e-54f6cebed7e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.314288] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e50f540a-dc3e-4049-a440-bf5d9722780a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.321759] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65766bfa-f7a6-47c6-a8b9-486faf6a2bcd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.338579] env[61768]: DEBUG nova.compute.provider_tree [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 940.364609] env[61768]: DEBUG nova.compute.manager [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Received event network-changed-9681c697-a94e-4494-a310-1468fe766740 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.364998] env[61768]: DEBUG nova.compute.manager [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Refreshing instance network info cache due to event network-changed-9681c697-a94e-4494-a310-1468fe766740. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.365094] env[61768]: DEBUG oslo_concurrency.lockutils [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] Acquiring lock "refresh_cache-71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.374481] env[61768]: DEBUG nova.network.neutron [-] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.468617] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5233fe61-a03a-a261-d423-f4f455c9cd4c, 'name': SearchDatastore_Task, 'duration_secs': 0.009841} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.468617] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.468617] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] df570987-0144-4e3a-83db-a9538d3835f5/df570987-0144-4e3a-83db-a9538d3835f5.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 940.469258] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-af79cacb-4a6b-4076-8aee-63b323d133d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.480877] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 940.480877] env[61768]: value = "task-1229255" [ 940.480877] env[61768]: _type = "Task" [ 940.480877] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.480877] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e439b764-1703-4ff7-9a67-c8630b96352e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.502428] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229255, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.504724] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2af193ec-df12-4331-964a-c98f7df43f55 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.538164] env[61768]: DEBUG nova.compute.manager [req-e3a68b1f-cb84-4ff2-9fd0-6b394720e054 req-a01694f9-4ea0-4382-9216-6dedf605526a service nova] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Detach interface failed, port_id=51ac6b12-5631-4ce5-acc4-2f43dee834d0, reason: Instance fb344f81-9295-44a1-9bac-dd530c157e98 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 940.614444] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.614778] env[61768]: DEBUG nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Instance network_info: |[{"id": "9681c697-a94e-4494-a310-1468fe766740", "address": "fa:16:3e:c5:1d:3f", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9681c697-a9", "ovs_interfaceid": "9681c697-a94e-4494-a310-1468fe766740", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.615108] env[61768]: DEBUG oslo_concurrency.lockutils [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] Acquired lock "refresh_cache-71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.615298] env[61768]: DEBUG nova.network.neutron [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Refreshing network info cache for port 9681c697-a94e-4494-a310-1468fe766740 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 940.620014] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:1d:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9681c697-a94e-4494-a310-1468fe766740', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.629582] env[61768]: DEBUG oslo.service.loopingcall [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.631518] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 940.631774] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-91b394c9-1734-40d6-bcd8-b11c97e27402 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.659178] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.659178] env[61768]: value = "task-1229256" [ 940.659178] env[61768]: _type = "Task" [ 940.659178] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.669854] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229256, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.844024] env[61768]: DEBUG nova.scheduler.client.report [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 940.879708] env[61768]: INFO nova.compute.manager [-] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Took 1.29 seconds to deallocate network for instance. [ 940.993929] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229255, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463986} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.994365] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] df570987-0144-4e3a-83db-a9538d3835f5/df570987-0144-4e3a-83db-a9538d3835f5.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 940.994622] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 940.995061] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ff6f394-ce52-4d12-a395-d41e1a1a2256 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.003308] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 941.003308] env[61768]: value = "task-1229257" [ 941.003308] env[61768]: _type = "Task" [ 941.003308] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.014500] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229257, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.020024] env[61768]: DEBUG nova.network.neutron [-] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.172065] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229256, 'name': CreateVM_Task, 'duration_secs': 0.404611} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.174598] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 941.175481] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.175874] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.176271] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 941.176850] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0816d12-c9fb-4784-8d81-bbdcb49ff6a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.182468] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 941.182468] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a17585-d34a-18ac-d09f-38ec73d7c23f" [ 941.182468] env[61768]: _type = "Task" [ 941.182468] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.192541] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a17585-d34a-18ac-d09f-38ec73d7c23f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.348221] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.750s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.352027] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.780s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.352027] env[61768]: DEBUG nova.objects.instance [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'resources' on Instance uuid 62cddff5-d499-4ef3-869d-3cdb0328640d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 941.377721] env[61768]: INFO nova.scheduler.client.report [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Deleted allocations for instance ecd08348-b22f-49c5-b13f-b187506e38d6 [ 941.389852] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.460678] env[61768]: DEBUG nova.network.neutron [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Updated VIF entry in instance network info cache for port 9681c697-a94e-4494-a310-1468fe766740. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 941.462901] env[61768]: DEBUG nova.network.neutron [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Updating instance_info_cache with network_info: [{"id": "9681c697-a94e-4494-a310-1468fe766740", "address": "fa:16:3e:c5:1d:3f", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9681c697-a9", "ovs_interfaceid": "9681c697-a94e-4494-a310-1468fe766740", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.514988] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229257, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.05943} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.515448] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.516378] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e060b0d3-f94a-4e04-acf6-00b482fae828 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.530018] env[61768]: INFO nova.compute.manager [-] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Took 1.43 seconds to deallocate network for instance. [ 941.541890] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] df570987-0144-4e3a-83db-a9538d3835f5/df570987-0144-4e3a-83db-a9538d3835f5.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.544056] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28423840-b8a2-42ce-9c51-ca51632b9d56 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.569512] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 941.569512] env[61768]: value = "task-1229258" [ 941.569512] env[61768]: _type = "Task" [ 941.569512] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.578277] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229258, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.694630] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a17585-d34a-18ac-d09f-38ec73d7c23f, 'name': SearchDatastore_Task, 'duration_secs': 0.008858} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.694958] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.695221] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.695472] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.695630] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.695820] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.696854] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-008fd7ae-4ca6-4370-af06-449519eba878 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.705936] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.705936] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 941.706451] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82ab812e-d327-48e6-ae54-b558b9737085 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.712684] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 941.712684] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52191c40-f01b-c3b2-baa3-3c490ed4c9c5" [ 941.712684] env[61768]: _type = "Task" [ 941.712684] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.720505] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52191c40-f01b-c3b2-baa3-3c490ed4c9c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.886713] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a44c1b04-72d0-47a6-93ac-3de166021646 tempest-AttachInterfacesTestJSON-1318544374 tempest-AttachInterfacesTestJSON-1318544374-project-member] Lock "ecd08348-b22f-49c5-b13f-b187506e38d6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.910s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.968410] env[61768]: DEBUG oslo_concurrency.lockutils [req-65e483da-e979-4a9f-b976-96083f0c9082 req-671bc612-dd86-4bee-ba17-bda37bb45cdd service nova] Releasing lock "refresh_cache-71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.995259] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cbd0620-6255-43e7-8662-391a5c32d76e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.002407] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab58853a-98d0-465f-80c0-a503c3cb2daa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.034479] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1eeb9da-f987-4827-910e-b57f05630f08 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.042605] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6ac81c-20b6-40e8-bd93-a68f52653f0d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.057742] env[61768]: DEBUG nova.compute.provider_tree [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 942.066574] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.079395] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229258, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.222982] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52191c40-f01b-c3b2-baa3-3c490ed4c9c5, 'name': SearchDatastore_Task, 'duration_secs': 0.008613} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.223781] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92630995-6061-455f-9145-7d7065677b4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.229486] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 942.229486] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525cda21-2159-69ea-d2dd-667a1524e42e" [ 942.229486] env[61768]: _type = "Task" [ 942.229486] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.237234] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525cda21-2159-69ea-d2dd-667a1524e42e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.393403] env[61768]: DEBUG nova.compute.manager [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Received event network-vif-deleted-bebae6b0-1f32-43c0-8d08-84e5239ed501 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.393621] env[61768]: DEBUG nova.compute.manager [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Received event network-changed-edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.393793] env[61768]: DEBUG nova.compute.manager [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Refreshing instance network info cache due to event network-changed-edd83e46-b8b7-4276-b563-16d34bb22e9b. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 942.394237] env[61768]: DEBUG oslo_concurrency.lockutils [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] Acquiring lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.394391] env[61768]: DEBUG oslo_concurrency.lockutils [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] Acquired lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.394571] env[61768]: DEBUG nova.network.neutron [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Refreshing network info cache for port edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 942.561086] env[61768]: DEBUG nova.scheduler.client.report [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.582681] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229258, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.739888] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525cda21-2159-69ea-d2dd-667a1524e42e, 'name': SearchDatastore_Task, 'duration_secs': 0.009207} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.740270] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.740550] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98/71ce99fa-3d9d-4ff6-b06a-2e27d575ab98.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 942.740805] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-812e0a1e-51a0-43a8-87d7-4514a859a938 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.749618] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 942.749618] env[61768]: value = "task-1229259" [ 942.749618] env[61768]: _type = "Task" [ 942.749618] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.756642] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.838307] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.838558] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.067140] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.716s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.073286] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.684s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.073508] env[61768]: DEBUG nova.objects.instance [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lazy-loading 'resources' on Instance uuid fb344f81-9295-44a1-9bac-dd530c157e98 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.097587] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229258, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.102312] env[61768]: INFO nova.scheduler.client.report [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted allocations for instance 62cddff5-d499-4ef3-869d-3cdb0328640d [ 943.183136] env[61768]: DEBUG nova.network.neutron [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updated VIF entry in instance network info cache for port edd83e46-b8b7-4276-b563-16d34bb22e9b. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 943.183516] env[61768]: DEBUG nova.network.neutron [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updating instance_info_cache with network_info: [{"id": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "address": "fa:16:3e:a7:9e:51", "network": {"id": "07321fe6-3bc8-41ee-9485-7873816a3b2e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-669074208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.201", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64f5133fd59844859b8e01d31f3320ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedd83e46-b8", "ovs_interfaceid": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.257947] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229259, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463263} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.258787] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98/71ce99fa-3d9d-4ff6-b06a-2e27d575ab98.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 943.259050] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 943.259344] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e8bb262d-5d1f-4673-b372-c4392a85c5cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.266549] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 943.266549] env[61768]: value = "task-1229260" [ 943.266549] env[61768]: _type = "Task" [ 943.266549] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.274257] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229260, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.340639] env[61768]: DEBUG nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 943.591718] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229258, 'name': ReconfigVM_Task, 'duration_secs': 1.997412} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.592066] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Reconfigured VM instance instance-0000005f to attach disk [datastore2] df570987-0144-4e3a-83db-a9538d3835f5/df570987-0144-4e3a-83db-a9538d3835f5.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 943.592738] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a1cf4c95-ad75-4701-9b37-a824445ca4a0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.603739] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 943.603739] env[61768]: value = "task-1229261" [ 943.603739] env[61768]: _type = "Task" [ 943.603739] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.617460] env[61768]: DEBUG nova.compute.manager [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Received event network-changed-edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 943.617718] env[61768]: DEBUG nova.compute.manager [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Refreshing instance network info cache due to event network-changed-edd83e46-b8b7-4276-b563-16d34bb22e9b. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 943.617867] env[61768]: DEBUG oslo_concurrency.lockutils [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] Acquiring lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.618444] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229261, 'name': Rename_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.618965] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a3727be8-3b82-4646-88c6-bec7aee2de5b tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "62cddff5-d499-4ef3-869d-3cdb0328640d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.403s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 943.687079] env[61768]: DEBUG oslo_concurrency.lockutils [req-f94fbd27-cc82-439b-a2f7-049bee679a87 req-0c5eeeb8-b108-4e2b-b632-c20b0770ba2a service nova] Releasing lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.687521] env[61768]: DEBUG oslo_concurrency.lockutils [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] Acquired lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.687723] env[61768]: DEBUG nova.network.neutron [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Refreshing network info cache for port edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 943.747762] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a0b1694-014b-46af-a97b-7b13a5dd0a42 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.755698] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc2c3e5-f5f0-42a1-9af7-ef6f77ef9003 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.789984] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f584fd-aadd-4162-a950-c00a43f4069a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.797057] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229260, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065715} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.799054] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.799823] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0206b7d9-aeac-44be-b6b0-0c34a38c6510 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.802983] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fee2fa72-5b2e-4691-ac7c-53e535dc0487 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.826736] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98/71ce99fa-3d9d-4ff6-b06a-2e27d575ab98.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.834982] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ff40f89-e993-4c53-a4d7-c8aa9be93790 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.849384] env[61768]: DEBUG nova.compute.provider_tree [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 943.858524] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 943.858524] env[61768]: value = "task-1229262" [ 943.858524] env[61768]: _type = "Task" [ 943.858524] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.870605] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229262, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.871512] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.112860] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229261, 'name': Rename_Task, 'duration_secs': 0.488828} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.112860] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 944.113032] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4eb5d01b-1ce5-43c4-bf7c-f12b5b4dabba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.119499] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 944.119499] env[61768]: value = "task-1229263" [ 944.119499] env[61768]: _type = "Task" [ 944.119499] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.127830] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229263, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.353928] env[61768]: DEBUG nova.scheduler.client.report [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.372046] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229262, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.632253] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229263, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.739040] env[61768]: DEBUG nova.network.neutron [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updated VIF entry in instance network info cache for port edd83e46-b8b7-4276-b563-16d34bb22e9b. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 944.739040] env[61768]: DEBUG nova.network.neutron [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updating instance_info_cache with network_info: [{"id": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "address": "fa:16:3e:a7:9e:51", "network": {"id": "07321fe6-3bc8-41ee-9485-7873816a3b2e", "bridge": "br-int", "label": "tempest-FloatingIPsAssociationNegativeTestJSON-669074208-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "64f5133fd59844859b8e01d31f3320ab", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4fe416-47a6-4542-b59d-8c71ab4d6503", "external-id": "nsx-vlan-transportzone-369", "segmentation_id": 369, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapedd83e46-b8", "ovs_interfaceid": "edd83e46-b8b7-4276-b563-16d34bb22e9b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.862020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.862020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.795s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.862020] env[61768]: DEBUG nova.objects.instance [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'resources' on Instance uuid 739faae0-5811-4f88-b56b-1350c0f7b8be {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 944.874326] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229262, 'name': ReconfigVM_Task, 'duration_secs': 0.763361} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.875608] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98/71ce99fa-3d9d-4ff6-b06a-2e27d575ab98.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.876775] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-29703cf5-b62e-4ee0-8c1e-dcab845bdcbc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.884034] env[61768]: INFO nova.scheduler.client.report [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Deleted allocations for instance fb344f81-9295-44a1-9bac-dd530c157e98 [ 944.893160] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 944.893160] env[61768]: value = "task-1229264" [ 944.893160] env[61768]: _type = "Task" [ 944.893160] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.902386] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229264, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.986367] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "1158fbc3-f17b-44f7-847f-01bdf328a74a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.986639] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.987174] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "1158fbc3-f17b-44f7-847f-01bdf328a74a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.987403] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.987599] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.990930] env[61768]: INFO nova.compute.manager [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Terminating instance [ 944.993083] env[61768]: DEBUG nova.compute.manager [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 944.993558] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 944.994405] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9620643e-93c0-4cdb-84de-1610dea58c36 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.002648] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 945.003211] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7eec1b50-c398-4cd1-8d78-376db9648219 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.009767] env[61768]: DEBUG oslo_vmware.api [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 945.009767] env[61768]: value = "task-1229265" [ 945.009767] env[61768]: _type = "Task" [ 945.009767] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.170478] env[61768]: DEBUG oslo_vmware.api [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229265, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.170478] env[61768]: DEBUG oslo_vmware.api [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229263, 'name': PowerOnVM_Task, 'duration_secs': 0.755396} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.170478] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 945.170478] env[61768]: INFO nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Took 9.46 seconds to spawn the instance on the hypervisor. [ 945.170478] env[61768]: DEBUG nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.170478] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce702ed-9c81-41db-a9b6-3d68679802c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.170478] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "d54c94a6-0a53-47a8-b024-1a05439da837" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.170478] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "d54c94a6-0a53-47a8-b024-1a05439da837" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.170478] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "d54c94a6-0a53-47a8-b024-1a05439da837-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.170478] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "d54c94a6-0a53-47a8-b024-1a05439da837-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.170478] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "d54c94a6-0a53-47a8-b024-1a05439da837-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.170478] env[61768]: INFO nova.compute.manager [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Terminating instance [ 945.170478] env[61768]: DEBUG nova.compute.manager [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 945.170478] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 945.170478] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9372573c-e96c-4b3e-bf64-63728b61b553 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.173408] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 945.173820] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0da28cf6-4e1c-4e80-90e8-a5f2c34d30d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.182744] env[61768]: DEBUG oslo_vmware.api [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 945.182744] env[61768]: value = "task-1229266" [ 945.182744] env[61768]: _type = "Task" [ 945.182744] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.192114] env[61768]: DEBUG oslo_vmware.api [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229266, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.246337] env[61768]: DEBUG oslo_concurrency.lockutils [req-99a38f42-d1b8-405f-8023-10fd7d797bb5 req-24c860ef-8f56-429c-9030-ae115b25f470 service nova] Releasing lock "refresh_cache-1158fbc3-f17b-44f7-847f-01bdf328a74a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.397308] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1195ec58-2ecd-420a-a41a-ecac9f3512f8 tempest-ServerDiskConfigTestJSON-2059651609 tempest-ServerDiskConfigTestJSON-2059651609-project-member] Lock "fb344f81-9295-44a1-9bac-dd530c157e98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.024s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.406697] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229264, 'name': Rename_Task, 'duration_secs': 0.177994} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.407278] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 945.408553] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d5dc9f1-e917-47bb-9e61-862062ae7898 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.417389] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 945.417389] env[61768]: value = "task-1229267" [ 945.417389] env[61768]: _type = "Task" [ 945.417389] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.428660] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229267, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.523401] env[61768]: DEBUG oslo_vmware.api [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229265, 'name': PowerOffVM_Task, 'duration_secs': 0.226735} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.523401] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 945.523401] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 945.523672] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-acbe1cd1-0c4b-4299-878f-f2a8a21fcc27 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.527537] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f9db787-be2b-4e31-973b-b4f7d3b39970 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.534602] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536f4a55-a5bb-4e2d-bdfe-c4b4e0c6dbac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.565612] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c543337-da92-482f-836a-c7daaeb7dd3c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.573269] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef32579-9343-4ba8-aa8a-cc78c09aa0c8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.588050] env[61768]: DEBUG nova.compute.provider_tree [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 945.592566] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 945.592566] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 945.592566] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Deleting the datastore file [datastore2] 1158fbc3-f17b-44f7-847f-01bdf328a74a {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.595507] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a75668ae-af8c-40d8-9440-df9546395441 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.598542] env[61768]: DEBUG oslo_vmware.api [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for the task: (returnval){ [ 945.598542] env[61768]: value = "task-1229269" [ 945.598542] env[61768]: _type = "Task" [ 945.598542] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.606862] env[61768]: DEBUG oslo_vmware.api [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229269, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.659888] env[61768]: INFO nova.compute.manager [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Took 20.01 seconds to build instance. [ 945.693113] env[61768]: DEBUG oslo_vmware.api [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229266, 'name': PowerOffVM_Task, 'duration_secs': 0.277621} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.693384] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 945.693568] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 945.693817] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dbee2562-2f08-4aaf-a5f6-5c48771eb682 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.755449] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 945.755690] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 945.755888] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleting the datastore file [datastore1] d54c94a6-0a53-47a8-b024-1a05439da837 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 945.756176] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ddf93bc7-5c08-433c-a974-26ff01f268e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.763893] env[61768]: DEBUG oslo_vmware.api [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 945.763893] env[61768]: value = "task-1229271" [ 945.763893] env[61768]: _type = "Task" [ 945.763893] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.772506] env[61768]: DEBUG oslo_vmware.api [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229271, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.928230] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229267, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.091516] env[61768]: DEBUG nova.scheduler.client.report [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 946.109551] env[61768]: DEBUG oslo_vmware.api [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Task: {'id': task-1229269, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124432} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.109888] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.110764] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 946.110764] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 946.110764] env[61768]: INFO nova.compute.manager [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Took 1.12 seconds to destroy the instance on the hypervisor. [ 946.110990] env[61768]: DEBUG oslo.service.loopingcall [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.111236] env[61768]: DEBUG nova.compute.manager [-] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.111789] env[61768]: DEBUG nova.network.neutron [-] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 946.162402] env[61768]: DEBUG oslo_concurrency.lockutils [None req-092001a9-455b-4e10-b2d4-086b3c26f6a5 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "df570987-0144-4e3a-83db-a9538d3835f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.522s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.283177] env[61768]: DEBUG oslo_vmware.api [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229271, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.317738} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.283581] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.283959] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 946.284506] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 946.284984] env[61768]: INFO nova.compute.manager [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Took 1.12 seconds to destroy the instance on the hypervisor. [ 946.285500] env[61768]: DEBUG oslo.service.loopingcall [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.285805] env[61768]: DEBUG nova.compute.manager [-] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 946.285941] env[61768]: DEBUG nova.network.neutron [-] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 946.433020] env[61768]: DEBUG oslo_vmware.api [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229267, 'name': PowerOnVM_Task, 'duration_secs': 0.705499} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.433020] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 946.433020] env[61768]: INFO nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Took 8.46 seconds to spawn the instance on the hypervisor. [ 946.433020] env[61768]: DEBUG nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 946.433914] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10c7202-58c4-42fa-a85a-e1d7165f17a2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.467703] env[61768]: DEBUG nova.compute.manager [req-4eaaee03-6ebb-4336-94a2-4353dee6187f req-c548aa81-6205-42fd-ba9f-700e205be248 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Received event network-vif-deleted-edd83e46-b8b7-4276-b563-16d34bb22e9b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.467807] env[61768]: INFO nova.compute.manager [req-4eaaee03-6ebb-4336-94a2-4353dee6187f req-c548aa81-6205-42fd-ba9f-700e205be248 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Neutron deleted interface edd83e46-b8b7-4276-b563-16d34bb22e9b; detaching it from the instance and deleting it from the info cache [ 946.467982] env[61768]: DEBUG nova.network.neutron [req-4eaaee03-6ebb-4336-94a2-4353dee6187f req-c548aa81-6205-42fd-ba9f-700e205be248 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.599911] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.736s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.601163] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.730s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.603428] env[61768]: INFO nova.compute.claims [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 946.642538] env[61768]: INFO nova.scheduler.client.report [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted allocations for instance 739faae0-5811-4f88-b56b-1350c0f7b8be [ 946.937851] env[61768]: DEBUG nova.network.neutron [-] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.960685] env[61768]: DEBUG nova.compute.manager [req-0695004b-d702-4675-884c-4bf6d25ea052 req-734e9e0f-a0df-4531-b5ec-69952b7e1453 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Received event network-vif-deleted-7660d6d5-91ae-41e4-ba9f-1c09038a38c1 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.961043] env[61768]: INFO nova.compute.manager [req-0695004b-d702-4675-884c-4bf6d25ea052 req-734e9e0f-a0df-4531-b5ec-69952b7e1453 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Neutron deleted interface 7660d6d5-91ae-41e4-ba9f-1c09038a38c1; detaching it from the instance and deleting it from the info cache [ 946.961276] env[61768]: DEBUG nova.network.neutron [req-0695004b-d702-4675-884c-4bf6d25ea052 req-734e9e0f-a0df-4531-b5ec-69952b7e1453 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.966980] env[61768]: INFO nova.compute.manager [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Took 20.54 seconds to build instance. [ 946.971475] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9cf64dd3-bbb5-4fa4-a924-7c245c6ade37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.978203] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.978445] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.988702] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a7b3256-90d7-4413-83b0-75be9d3f97e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.018681] env[61768]: DEBUG nova.compute.manager [req-4eaaee03-6ebb-4336-94a2-4353dee6187f req-c548aa81-6205-42fd-ba9f-700e205be248 service nova] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Detach interface failed, port_id=edd83e46-b8b7-4276-b563-16d34bb22e9b, reason: Instance 1158fbc3-f17b-44f7-847f-01bdf328a74a could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 947.152924] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0c2ee43d-8383-48dd-849c-f997b3f391c3 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "739faae0-5811-4f88-b56b-1350c0f7b8be" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.197s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.311234] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "df570987-0144-4e3a-83db-a9538d3835f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.312025] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "df570987-0144-4e3a-83db-a9538d3835f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.312292] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "df570987-0144-4e3a-83db-a9538d3835f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.312923] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "df570987-0144-4e3a-83db-a9538d3835f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.313137] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "df570987-0144-4e3a-83db-a9538d3835f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.317531] env[61768]: INFO nova.compute.manager [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Terminating instance [ 947.317531] env[61768]: DEBUG nova.compute.manager [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 947.317531] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 947.318571] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4bf405e-0fd2-42fd-a903-8ec07970d8fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.327114] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 947.327372] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b136c28f-f743-4d38-8aa4-74bacbfc8318 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.334861] env[61768]: DEBUG oslo_vmware.api [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 947.334861] env[61768]: value = "task-1229272" [ 947.334861] env[61768]: _type = "Task" [ 947.334861] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.343699] env[61768]: DEBUG oslo_vmware.api [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229272, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.423429] env[61768]: DEBUG nova.network.neutron [-] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.440862] env[61768]: INFO nova.compute.manager [-] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Took 1.33 seconds to deallocate network for instance. [ 947.465431] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0394dbe8-36bc-4446-a2fd-699810e8f8da tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.052s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.467041] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca169ebb-de11-4b43-b730-5399ed1c1ae5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.477910] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3afd859-c92f-4369-80dc-c7eab051d82c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.493244] env[61768]: DEBUG nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 947.515613] env[61768]: DEBUG nova.compute.manager [req-0695004b-d702-4675-884c-4bf6d25ea052 req-734e9e0f-a0df-4531-b5ec-69952b7e1453 service nova] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Detach interface failed, port_id=7660d6d5-91ae-41e4-ba9f-1c09038a38c1, reason: Instance d54c94a6-0a53-47a8-b024-1a05439da837 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 947.763905] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e09381d4-f7a3-42d6-8cf3-6aa20a34996d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.772334] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69db5c6e-46ac-4be6-958b-1e2b9d20ef68 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.802876] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-393e90cc-25c3-470e-8a11-bd3112f41530 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.813095] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14587a44-9e14-4a28-ba32-b7b6bbe1f8bc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.827899] env[61768]: DEBUG nova.compute.provider_tree [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.848280] env[61768]: DEBUG oslo_vmware.api [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229272, 'name': PowerOffVM_Task, 'duration_secs': 0.319717} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.848753] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 947.848860] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 947.849041] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6ff5d95-d9cc-4abd-bbf8-07131a72c456 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.926218] env[61768]: INFO nova.compute.manager [-] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Took 1.64 seconds to deallocate network for instance. [ 947.926614] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 947.926829] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 947.927111] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Deleting the datastore file [datastore2] df570987-0144-4e3a-83db-a9538d3835f5 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.928908] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7a3a44e5-ca87-4c25-9b64-540bc5af27be {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.937985] env[61768]: DEBUG oslo_vmware.api [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for the task: (returnval){ [ 947.937985] env[61768]: value = "task-1229274" [ 947.937985] env[61768]: _type = "Task" [ 947.937985] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.946783] env[61768]: DEBUG oslo_vmware.api [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229274, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.949392] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.013205] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.059301] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.059301] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.059532] env[61768]: DEBUG nova.compute.manager [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.060531] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66d3afb-c192-4011-9acc-4713b55dc8a8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.070798] env[61768]: DEBUG nova.compute.manager [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61768) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 948.073721] env[61768]: DEBUG nova.objects.instance [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'flavor' on Instance uuid 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.331609] env[61768]: DEBUG nova.scheduler.client.report [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.435932] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.454392] env[61768]: DEBUG oslo_vmware.api [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Task: {'id': task-1229274, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131169} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.454677] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 948.454867] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 948.455060] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 948.455761] env[61768]: INFO nova.compute.manager [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Took 1.14 seconds to destroy the instance on the hypervisor. [ 948.455761] env[61768]: DEBUG oslo.service.loopingcall [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 948.455761] env[61768]: DEBUG nova.compute.manager [-] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 948.455761] env[61768]: DEBUG nova.network.neutron [-] [instance: df570987-0144-4e3a-83db-a9538d3835f5] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 948.580119] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 948.580865] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5112dfe-4777-431c-b1bc-8859bc472c80 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.587729] env[61768]: DEBUG oslo_vmware.api [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 948.587729] env[61768]: value = "task-1229275" [ 948.587729] env[61768]: _type = "Task" [ 948.587729] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.625422] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 948.625681] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.840065] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.238s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.840204] env[61768]: DEBUG nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 948.843692] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.894s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.843933] env[61768]: DEBUG nova.objects.instance [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lazy-loading 'resources' on Instance uuid 1158fbc3-f17b-44f7-847f-01bdf328a74a {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.986570] env[61768]: DEBUG nova.compute.manager [req-9c031942-2ba5-443a-91d4-6c9fa1101a7e req-3e533a08-309d-4aa5-b347-ea36f23c18d5 service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Received event network-vif-deleted-94ba3a4c-7b23-48e1-a23d-bef8d5eaa582 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 948.986570] env[61768]: INFO nova.compute.manager [req-9c031942-2ba5-443a-91d4-6c9fa1101a7e req-3e533a08-309d-4aa5-b347-ea36f23c18d5 service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Neutron deleted interface 94ba3a4c-7b23-48e1-a23d-bef8d5eaa582; detaching it from the instance and deleting it from the info cache [ 948.986570] env[61768]: DEBUG nova.network.neutron [req-9c031942-2ba5-443a-91d4-6c9fa1101a7e req-3e533a08-309d-4aa5-b347-ea36f23c18d5 service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.102202] env[61768]: DEBUG oslo_vmware.api [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229275, 'name': PowerOffVM_Task, 'duration_secs': 0.261981} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.102202] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 949.102202] env[61768]: DEBUG nova.compute.manager [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 949.102846] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4906e296-ceb8-4538-ad26-8b04520f5816 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.128095] env[61768]: DEBUG nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 949.286228] env[61768]: DEBUG nova.network.neutron [-] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 949.347534] env[61768]: DEBUG nova.compute.utils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 949.353314] env[61768]: DEBUG nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 949.354161] env[61768]: DEBUG nova.network.neutron [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 949.427030] env[61768]: DEBUG nova.policy [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c39490b7563b4b20abbe27606b42f58d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f844c9d775474fb69d9ea04406cc1632', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 949.490744] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14a6fe07-fad4-4033-9299-05990c5f8722 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.502215] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71192ad4-8460-4884-9d05-0e45191a15c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.532090] env[61768]: DEBUG nova.compute.manager [req-9c031942-2ba5-443a-91d4-6c9fa1101a7e req-3e533a08-309d-4aa5-b347-ea36f23c18d5 service nova] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Detach interface failed, port_id=94ba3a4c-7b23-48e1-a23d-bef8d5eaa582, reason: Instance df570987-0144-4e3a-83db-a9538d3835f5 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 949.533901] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f7ff36-d4f7-476a-aa58-e7d12e535522 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.540759] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9a386b-1982-4de8-b920-518e8960ed46 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.571575] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e9696d-765e-43bd-a00c-250d0f1c8061 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.580359] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97de0e30-004d-4b37-92c7-d48fd39b8ec7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.595738] env[61768]: DEBUG nova.compute.provider_tree [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.614830] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4590bdba-01ad-47d6-a8e1-c25f2a3fc2c5 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.555s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.650381] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.787920] env[61768]: INFO nova.compute.manager [-] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Took 1.33 seconds to deallocate network for instance. [ 949.853746] env[61768]: DEBUG nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.102024] env[61768]: DEBUG nova.scheduler.client.report [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 950.113730] env[61768]: DEBUG nova.network.neutron [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Successfully created port: 3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.296619] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.612428] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.767s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.614026] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.600s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.615364] env[61768]: INFO nova.compute.claims [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 950.634073] env[61768]: INFO nova.scheduler.client.report [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Deleted allocations for instance 1158fbc3-f17b-44f7-847f-01bdf328a74a [ 950.814898] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "49723fff-d1ce-476f-ac23-bd2472762d94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.815214] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "49723fff-d1ce-476f-ac23-bd2472762d94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.868787] env[61768]: DEBUG nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 950.896462] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.896786] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.896956] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.898068] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.898369] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.898937] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.898937] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.899096] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.899800] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.899800] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.899800] env[61768]: DEBUG nova.virt.hardware [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.900977] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbea1a3-2ad1-4189-a5b3-a3c1ea90025f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.911756] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07565a8b-bb04-492d-835a-0b761bf32bf6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.048329] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.049781] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.050171] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.050549] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.050549] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.054616] env[61768]: INFO nova.compute.manager [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Terminating instance [ 951.058922] env[61768]: DEBUG nova.compute.manager [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 951.059219] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 951.060489] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34ec1926-0149-4ac5-9107-706b54d02188 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.069683] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 951.070237] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49fd8fb9-6430-4dd6-94a8-ef38ea2a2589 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.137107] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 951.137530] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 951.137530] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleting the datastore file [datastore2] 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 951.137784] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8c8690b3-a040-4eaf-afbe-39fcbd21c5ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.145363] env[61768]: DEBUG oslo_vmware.api [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 951.145363] env[61768]: value = "task-1229277" [ 951.145363] env[61768]: _type = "Task" [ 951.145363] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.145804] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ef3ea9d8-6b0b-43b2-8dfd-07a16eb96bb3 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535 tempest-FloatingIPsAssociationNegativeTestJSON-1054198535-project-member] Lock "1158fbc3-f17b-44f7-847f-01bdf328a74a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.159s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.156091] env[61768]: DEBUG oslo_vmware.api [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229277, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.321611] env[61768]: DEBUG nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 951.663605] env[61768]: DEBUG oslo_vmware.api [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229277, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126444} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.664862] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 951.666519] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 951.666865] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 951.667562] env[61768]: INFO nova.compute.manager [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Took 0.61 seconds to destroy the instance on the hypervisor. [ 951.669184] env[61768]: DEBUG oslo.service.loopingcall [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.673197] env[61768]: DEBUG nova.compute.manager [-] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 951.681190] env[61768]: DEBUG nova.network.neutron [-] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 951.681190] env[61768]: DEBUG nova.compute.manager [req-58b9c009-66e2-4b21-bf5b-95d1058eda8b req-ec0de93f-7db5-4367-bde2-4e952e134b7c service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Received event network-vif-plugged-3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 951.681190] env[61768]: DEBUG oslo_concurrency.lockutils [req-58b9c009-66e2-4b21-bf5b-95d1058eda8b req-ec0de93f-7db5-4367-bde2-4e952e134b7c service nova] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.681190] env[61768]: DEBUG oslo_concurrency.lockutils [req-58b9c009-66e2-4b21-bf5b-95d1058eda8b req-ec0de93f-7db5-4367-bde2-4e952e134b7c service nova] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.681190] env[61768]: DEBUG oslo_concurrency.lockutils [req-58b9c009-66e2-4b21-bf5b-95d1058eda8b req-ec0de93f-7db5-4367-bde2-4e952e134b7c service nova] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.681190] env[61768]: DEBUG nova.compute.manager [req-58b9c009-66e2-4b21-bf5b-95d1058eda8b req-ec0de93f-7db5-4367-bde2-4e952e134b7c service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] No waiting events found dispatching network-vif-plugged-3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 951.681190] env[61768]: WARNING nova.compute.manager [req-58b9c009-66e2-4b21-bf5b-95d1058eda8b req-ec0de93f-7db5-4367-bde2-4e952e134b7c service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Received unexpected event network-vif-plugged-3ceeacd4-686d-4a24-bafc-328b8bd344ed for instance with vm_state building and task_state spawning. [ 951.791332] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7e1f734-a653-4f3b-aceb-ffbd77006382 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.802869] env[61768]: DEBUG nova.network.neutron [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Successfully updated port: 3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.811260] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37aed6d6-eac0-4078-b033-8effd07bde18 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.852289] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35089904-b35f-4791-b405-d3ce283cf62d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.863121] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cce8b1c-87f7-43c0-87b3-6f91d535a988 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.867522] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.877906] env[61768]: DEBUG nova.compute.provider_tree [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.223637] env[61768]: DEBUG nova.compute.manager [req-a0dc5dc5-a660-43ae-9bd4-182b7ef86457 req-db049371-0032-4067-a244-fc86295e6b08 service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Received event network-vif-deleted-9681c697-a94e-4494-a310-1468fe766740 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.223884] env[61768]: INFO nova.compute.manager [req-a0dc5dc5-a660-43ae-9bd4-182b7ef86457 req-db049371-0032-4067-a244-fc86295e6b08 service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Neutron deleted interface 9681c697-a94e-4494-a310-1468fe766740; detaching it from the instance and deleting it from the info cache [ 952.224033] env[61768]: DEBUG nova.network.neutron [req-a0dc5dc5-a660-43ae-9bd4-182b7ef86457 req-db049371-0032-4067-a244-fc86295e6b08 service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.308777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.308960] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.309131] env[61768]: DEBUG nova.network.neutron [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 952.359689] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 952.360016] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 952.383153] env[61768]: DEBUG nova.scheduler.client.report [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 952.689685] env[61768]: DEBUG nova.network.neutron [-] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.726687] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fbd20e3e-bc9a-40f9-8277-2f66e6c1c99d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.736820] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f7eda6d-4003-4f4d-8753-0bdc7271a39c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.763718] env[61768]: DEBUG nova.compute.manager [req-a0dc5dc5-a660-43ae-9bd4-182b7ef86457 req-db049371-0032-4067-a244-fc86295e6b08 service nova] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Detach interface failed, port_id=9681c697-a94e-4494-a310-1468fe766740, reason: Instance 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 952.854223] env[61768]: DEBUG nova.network.neutron [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 952.866183] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 952.866341] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 952.889160] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.276s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.889757] env[61768]: DEBUG nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 952.898732] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.463s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.898928] env[61768]: DEBUG nova.objects.instance [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'resources' on Instance uuid d54c94a6-0a53-47a8-b024-1a05439da837 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.111292] env[61768]: DEBUG nova.network.neutron [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updating instance_info_cache with network_info: [{"id": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "address": "fa:16:3e:fb:f6:d9", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ceeacd4-68", "ovs_interfaceid": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.192274] env[61768]: INFO nova.compute.manager [-] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Took 1.51 seconds to deallocate network for instance. [ 953.407885] env[61768]: DEBUG nova.compute.utils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 953.407885] env[61768]: DEBUG nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 953.407885] env[61768]: DEBUG nova.network.neutron [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 953.485369] env[61768]: DEBUG nova.policy [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dc4b77aedab34b96a303a133bc771e8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa61ecc36bc14769a2fc909ee95716ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 953.617929] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.618784] env[61768]: DEBUG nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Instance network_info: |[{"id": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "address": "fa:16:3e:fb:f6:d9", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ceeacd4-68", "ovs_interfaceid": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 953.621025] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:f6:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ccc845e3-654b-43c6-acea-dde1084f0ad0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3ceeacd4-686d-4a24-bafc-328b8bd344ed', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 953.630616] env[61768]: DEBUG oslo.service.loopingcall [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 953.630894] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 953.631241] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73067cb8-f7a8-4c61-bd1c-4dfbf5fa8d01 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.651968] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c727eaf8-6e14-4d71-a47c-647989b798dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.661837] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd6cc30-cf6b-4235-8791-08768e693bc1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.665508] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 953.665508] env[61768]: value = "task-1229278" [ 953.665508] env[61768]: _type = "Task" [ 953.665508] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.673521] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "750bc93c-65aa-4afd-a07b-cbeda0acae24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.673824] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.701204] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.703437] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.703437] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquired lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.703539] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Forcefully refreshing network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 953.705492] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3043ce28-8a56-4f9a-952c-3643d3966e98 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.720381] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfe7e240-e40e-4636-8fae-a5eb9230311d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.727371] env[61768]: DEBUG nova.compute.manager [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Received event network-changed-3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 953.727571] env[61768]: DEBUG nova.compute.manager [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Refreshing instance network info cache due to event network-changed-3ceeacd4-686d-4a24-bafc-328b8bd344ed. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 953.727789] env[61768]: DEBUG oslo_concurrency.lockutils [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] Acquiring lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.727940] env[61768]: DEBUG oslo_concurrency.lockutils [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] Acquired lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.728132] env[61768]: DEBUG nova.network.neutron [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Refreshing network info cache for port 3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 953.738308] env[61768]: DEBUG nova.compute.provider_tree [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.911348] env[61768]: DEBUG nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 953.916243] env[61768]: DEBUG nova.network.neutron [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Successfully created port: 8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.177643] env[61768]: DEBUG nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.180569] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229278, 'name': CreateVM_Task, 'duration_secs': 0.388455} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.180949] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 954.181817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.181997] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.182338] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 954.182603] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cdcf0116-8cca-4863-bef0-6b8e52eb34fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.189098] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 954.189098] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c876e7-2642-7566-4722-2fbe95d24673" [ 954.189098] env[61768]: _type = "Task" [ 954.189098] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.196594] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c876e7-2642-7566-4722-2fbe95d24673, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.241173] env[61768]: DEBUG nova.scheduler.client.report [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.445716] env[61768]: DEBUG nova.network.neutron [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updated VIF entry in instance network info cache for port 3ceeacd4-686d-4a24-bafc-328b8bd344ed. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 954.446101] env[61768]: DEBUG nova.network.neutron [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updating instance_info_cache with network_info: [{"id": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "address": "fa:16:3e:fb:f6:d9", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ceeacd4-68", "ovs_interfaceid": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.698884] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c876e7-2642-7566-4722-2fbe95d24673, 'name': SearchDatastore_Task, 'duration_secs': 0.009631} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.699081] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.699323] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 954.699568] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.699725] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.699913] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 954.700191] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4459c13-6db1-4119-aafa-74d3e77b5428 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.702591] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.708612] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 954.708795] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 954.709474] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc7262d8-f281-4075-b006-268e0cd66123 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.714118] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 954.714118] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d9b331-d36d-534e-8aa9-3bb8ab8df2b3" [ 954.714118] env[61768]: _type = "Task" [ 954.714118] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.721165] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d9b331-d36d-534e-8aa9-3bb8ab8df2b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.750132] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.752368] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.102s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.753886] env[61768]: INFO nova.compute.claims [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 954.768079] env[61768]: INFO nova.scheduler.client.report [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted allocations for instance d54c94a6-0a53-47a8-b024-1a05439da837 [ 954.920075] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Updating instance_info_cache with network_info: [{"id": "49a82793-2751-49db-b262-593b386ea68d", "address": "fa:16:3e:67:7c:ab", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap49a82793-27", "ovs_interfaceid": "49a82793-2751-49db-b262-593b386ea68d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.922500] env[61768]: DEBUG nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 954.947125] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 954.947430] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 954.947603] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 954.947798] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 954.947954] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 954.948127] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 954.948368] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 954.948541] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 954.948719] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 954.948890] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 954.949085] env[61768]: DEBUG nova.virt.hardware [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 954.949817] env[61768]: DEBUG oslo_concurrency.lockutils [req-21c913ae-0154-490e-a3d3-5328f146589b req-69eada17-6d9a-48d3-b934-f2cd157b993e service nova] Releasing lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.950777] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba30890-1a76-4e61-a2f3-4846b8f38d69 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.958869] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0a728c0-eb0e-4c9d-9779-aefc401c9c5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.225061] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52d9b331-d36d-534e-8aa9-3bb8ab8df2b3, 'name': SearchDatastore_Task, 'duration_secs': 0.008481} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.225364] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27d824a2-dc96-4e6c-af70-fcc60f7bfe1f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.230538] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 955.230538] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a4821-40b5-601b-72a1-04250afdd757" [ 955.230538] env[61768]: _type = "Task" [ 955.230538] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.238012] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a4821-40b5-601b-72a1-04250afdd757, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.274998] env[61768]: DEBUG oslo_concurrency.lockutils [None req-f8b97954-4484-487e-973d-c8540fd7b527 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "d54c94a6-0a53-47a8-b024-1a05439da837" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.120s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.425987] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Releasing lock "refresh_cache-2c85de5d-0111-40a6-a8bc-69c7eba0393d" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.426232] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Updated the network info_cache for instance {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 955.426441] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.426605] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.426757] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.426974] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.427174] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.427332] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.427566] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 955.427685] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.527106] env[61768]: DEBUG nova.network.neutron [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Successfully updated port: 8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 955.741246] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529a4821-40b5-601b-72a1-04250afdd757, 'name': SearchDatastore_Task, 'duration_secs': 0.009894} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.741591] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.741908] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] e1e90c98-5a65-473c-b508-b45fd93e31ad/e1e90c98-5a65-473c-b508-b45fd93e31ad.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 955.742235] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32a17f57-254f-452f-850c-d4f8e9a7e6e7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.748898] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 955.748898] env[61768]: value = "task-1229279" [ 955.748898] env[61768]: _type = "Task" [ 955.748898] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.757125] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229279, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.759176] env[61768]: DEBUG nova.compute.manager [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-vif-plugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.759400] env[61768]: DEBUG oslo_concurrency.lockutils [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.759616] env[61768]: DEBUG oslo_concurrency.lockutils [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.759789] env[61768]: DEBUG oslo_concurrency.lockutils [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.759966] env[61768]: DEBUG nova.compute.manager [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] No waiting events found dispatching network-vif-plugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.760159] env[61768]: WARNING nova.compute.manager [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received unexpected event network-vif-plugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 for instance with vm_state building and task_state spawning. [ 955.760332] env[61768]: DEBUG nova.compute.manager [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 955.760543] env[61768]: DEBUG nova.compute.manager [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing instance network info cache due to event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 955.760886] env[61768]: DEBUG oslo_concurrency.lockutils [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.761160] env[61768]: DEBUG oslo_concurrency.lockutils [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.761369] env[61768]: DEBUG nova.network.neutron [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing network info cache for port 8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 955.911073] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae030318-5e22-49a2-96cf-dbb951debd3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.920139] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03d72b53-a518-4d3c-b0f9-101eab1c3e20 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.931038] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.978562] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5b4110-4477-4d39-b4f7-e596380f4596 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.993025] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c9390c-ffcf-4c88-ac34-0b4383c9852a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.016342] env[61768]: DEBUG nova.compute.provider_tree [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.030665] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.260510] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229279, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.297580] env[61768]: DEBUG nova.network.neutron [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 956.376691] env[61768]: DEBUG nova.network.neutron [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.520086] env[61768]: DEBUG nova.scheduler.client.report [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 956.713766] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "8e0e13a7-eac5-4176-8536-2906d13c390e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.714088] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.760868] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229279, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544639} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.761142] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] e1e90c98-5a65-473c-b508-b45fd93e31ad/e1e90c98-5a65-473c-b508-b45fd93e31ad.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 956.761366] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 956.761646] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e949e94-7024-4ed6-9bd6-db2a5ff88e2c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.768236] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 956.768236] env[61768]: value = "task-1229280" [ 956.768236] env[61768]: _type = "Task" [ 956.768236] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.775424] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229280, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.879318] env[61768]: DEBUG oslo_concurrency.lockutils [req-8d656617-46ec-4a97-99d6-6a997d591ef1 req-70d8cedc-6a6b-42b5-83d7-e812ce71be0d service nova] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.879742] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.879910] env[61768]: DEBUG nova.network.neutron [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 957.025557] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.273s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.026107] env[61768]: DEBUG nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.028704] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.732s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.028942] env[61768]: DEBUG nova.objects.instance [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lazy-loading 'resources' on Instance uuid df570987-0144-4e3a-83db-a9538d3835f5 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 957.216913] env[61768]: INFO nova.compute.manager [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Detaching volume d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b [ 957.248738] env[61768]: INFO nova.virt.block_device [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Attempting to driver detach volume d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b from mountpoint /dev/sdb [ 957.248997] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 957.249224] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265490', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'name': 'volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e0e13a7-eac5-4176-8536-2906d13c390e', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'serial': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 957.250099] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd82414-df54-44c6-bf6f-9645e049a482 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.274430] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc5f96b-3a09-4744-ab37-f8bde6142066 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.282351] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229280, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067644} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.283536] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 957.284398] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7d963cb-cfc4-4a9f-86eb-ccaea535c786 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.287081] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f042165-587e-4b02-a258-0a3e7c73f6a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.308364] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] e1e90c98-5a65-473c-b508-b45fd93e31ad/e1e90c98-5a65-473c-b508-b45fd93e31ad.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 957.323381] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16e1b262-1f34-4ac7-acf5-fbac34570c94 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.338092] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84fd66ac-1efe-43e7-9380-8b2e5a40b732 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.354698] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] The volume has not been displaced from its original location: [datastore2] volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b/volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 957.359857] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 957.361160] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e01c44d-e26a-4e2c-8ba6-cd96fdc998cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.374815] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 957.374815] env[61768]: value = "task-1229281" [ 957.374815] env[61768]: _type = "Task" [ 957.374815] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.380456] env[61768]: DEBUG oslo_vmware.api [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 957.380456] env[61768]: value = "task-1229282" [ 957.380456] env[61768]: _type = "Task" [ 957.380456] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.384930] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229281, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.392848] env[61768]: DEBUG oslo_vmware.api [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229282, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.416165] env[61768]: DEBUG nova.network.neutron [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 957.531970] env[61768]: DEBUG nova.compute.utils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 957.533693] env[61768]: DEBUG nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 957.533903] env[61768]: DEBUG nova.network.neutron [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 957.572830] env[61768]: DEBUG nova.network.neutron [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.606366] env[61768]: DEBUG nova.policy [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fd5852bc16429aa2904b4351c8249d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7857b5217cc499ca8c9c3b8a5471db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 957.676841] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c586a83c-ac78-40e3-bed2-05190b8c00ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.684817] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29fa7665-1808-4601-ba4f-729439472b86 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.713735] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3255988-4ee1-45a2-8067-50ae3d05ae3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.720993] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98af59cf-5ccf-441b-a9fe-5d92162c433f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.735555] env[61768]: DEBUG nova.compute.provider_tree [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 957.884991] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229281, 'name': ReconfigVM_Task, 'duration_secs': 0.294647} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.892087] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfigured VM instance instance-00000061 to attach disk [datastore2] e1e90c98-5a65-473c-b508-b45fd93e31ad/e1e90c98-5a65-473c-b508-b45fd93e31ad.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.892087] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8f6c03fe-dafa-45ab-8e64-44e07d214666 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.896213] env[61768]: DEBUG oslo_vmware.api [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229282, 'name': ReconfigVM_Task, 'duration_secs': 0.227738} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.896871] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 957.902034] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 957.902034] env[61768]: value = "task-1229283" [ 957.902034] env[61768]: _type = "Task" [ 957.902034] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.902034] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-739827c3-2c15-4beb-8338-935427458bae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.913026] env[61768]: DEBUG nova.network.neutron [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Successfully created port: 52cbc403-960a-4e8e-ab52-b1ea176a6de1 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 957.925022] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229283, 'name': Rename_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.925022] env[61768]: DEBUG oslo_vmware.api [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 957.925022] env[61768]: value = "task-1229284" [ 957.925022] env[61768]: _type = "Task" [ 957.925022] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.932448] env[61768]: DEBUG oslo_vmware.api [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229284, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.039181] env[61768]: DEBUG nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.078074] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.078414] env[61768]: DEBUG nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Instance network_info: |[{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 958.078992] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:30:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37434b93-dfdc-4a3f-bf5a-9f2cbe25a754', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e676cc0-ee67-4e76-afe1-2dff2e254617', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.087098] env[61768]: DEBUG oslo.service.loopingcall [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.087586] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 958.087827] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d13f20ce-d139-48df-a3c2-575791ff65b9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.108354] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.108354] env[61768]: value = "task-1229285" [ 958.108354] env[61768]: _type = "Task" [ 958.108354] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.116285] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229285, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.238242] env[61768]: DEBUG nova.scheduler.client.report [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.421931] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229283, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.432816] env[61768]: DEBUG oslo_vmware.api [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229284, 'name': ReconfigVM_Task, 'duration_secs': 0.140845} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.433162] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265490', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'name': 'volume-d2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '8e0e13a7-eac5-4176-8536-2906d13c390e', 'attached_at': '', 'detached_at': '', 'volume_id': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b', 'serial': 'd2d48ade-0ab9-4e20-b6db-fb8adf5bce8b'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 958.619284] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229285, 'name': CreateVM_Task} progress is 25%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.742851] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.714s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.745809] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.878s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.747112] env[61768]: INFO nova.compute.claims [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 958.768388] env[61768]: INFO nova.scheduler.client.report [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Deleted allocations for instance df570987-0144-4e3a-83db-a9538d3835f5 [ 958.926096] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229283, 'name': Rename_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.981019] env[61768]: DEBUG nova.objects.instance [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'flavor' on Instance uuid 8e0e13a7-eac5-4176-8536-2906d13c390e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 959.050819] env[61768]: DEBUG nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.076510] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.076804] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.077091] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.077253] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.077498] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.077695] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.077978] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.078111] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.078352] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.078562] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.078761] env[61768]: DEBUG nova.virt.hardware [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.079702] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9c7fda-3052-44bf-856b-7a60bc448565 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.087779] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a09632c7-eb4b-4564-82f7-63193df852dc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.118490] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229285, 'name': CreateVM_Task} progress is 25%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.233591] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Acquiring lock "30f55051-63b3-47a9-83fb-945cdeb82574" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.233888] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "30f55051-63b3-47a9-83fb-945cdeb82574" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.234130] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Acquiring lock "30f55051-63b3-47a9-83fb-945cdeb82574-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.234329] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "30f55051-63b3-47a9-83fb-945cdeb82574-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.234513] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "30f55051-63b3-47a9-83fb-945cdeb82574-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.238873] env[61768]: INFO nova.compute.manager [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Terminating instance [ 959.242468] env[61768]: DEBUG nova.compute.manager [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 959.242468] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 959.242468] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-10a4a1b1-4ef9-4b6a-8d1b-de01568211e3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.252491] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 959.252491] env[61768]: value = "task-1229286" [ 959.252491] env[61768]: _type = "Task" [ 959.252491] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.262747] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.280824] env[61768]: DEBUG oslo_concurrency.lockutils [None req-3357ce32-8a33-4ff7-95ab-e6da2619a1c0 tempest-ServerAddressesTestJSON-1795000325 tempest-ServerAddressesTestJSON-1795000325-project-member] Lock "df570987-0144-4e3a-83db-a9538d3835f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.968s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.374208] env[61768]: DEBUG nova.compute.manager [req-4abb5cad-cdb7-4540-aaee-0b4220195dc0 req-cb90b0d4-2dc6-4582-8310-73fee466b385 service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Received event network-vif-plugged-52cbc403-960a-4e8e-ab52-b1ea176a6de1 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.374486] env[61768]: DEBUG oslo_concurrency.lockutils [req-4abb5cad-cdb7-4540-aaee-0b4220195dc0 req-cb90b0d4-2dc6-4582-8310-73fee466b385 service nova] Acquiring lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.374744] env[61768]: DEBUG oslo_concurrency.lockutils [req-4abb5cad-cdb7-4540-aaee-0b4220195dc0 req-cb90b0d4-2dc6-4582-8310-73fee466b385 service nova] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.375067] env[61768]: DEBUG oslo_concurrency.lockutils [req-4abb5cad-cdb7-4540-aaee-0b4220195dc0 req-cb90b0d4-2dc6-4582-8310-73fee466b385 service nova] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.375282] env[61768]: DEBUG nova.compute.manager [req-4abb5cad-cdb7-4540-aaee-0b4220195dc0 req-cb90b0d4-2dc6-4582-8310-73fee466b385 service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] No waiting events found dispatching network-vif-plugged-52cbc403-960a-4e8e-ab52-b1ea176a6de1 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 959.375500] env[61768]: WARNING nova.compute.manager [req-4abb5cad-cdb7-4540-aaee-0b4220195dc0 req-cb90b0d4-2dc6-4582-8310-73fee466b385 service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Received unexpected event network-vif-plugged-52cbc403-960a-4e8e-ab52-b1ea176a6de1 for instance with vm_state building and task_state spawning. [ 959.422627] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229283, 'name': Rename_Task, 'duration_secs': 1.162893} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.422932] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 959.423222] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8297466f-b4e7-4417-ad17-effdf05897f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.429325] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 959.429325] env[61768]: value = "task-1229287" [ 959.429325] env[61768]: _type = "Task" [ 959.429325] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.438440] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229287, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.478190] env[61768]: DEBUG nova.network.neutron [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Successfully updated port: 52cbc403-960a-4e8e-ab52-b1ea176a6de1 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 959.619699] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229285, 'name': CreateVM_Task, 'duration_secs': 1.269088} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.619874] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 959.620620] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.620804] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.621164] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 959.621426] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-836fdcbf-055b-44fb-9096-6d0eb95bf49b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.626116] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 959.626116] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524203f0-d2d4-38ba-f974-865ac61bcbf8" [ 959.626116] env[61768]: _type = "Task" [ 959.626116] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.633885] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524203f0-d2d4-38ba-f974-865ac61bcbf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.766723] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229286, 'name': PowerOffVM_Task, 'duration_secs': 0.171316} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.767030] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 959.767260] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 959.767489] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265498', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'name': 'volume-68c9b861-d548-4ed9-b618-463dfbd8cc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '30f55051-63b3-47a9-83fb-945cdeb82574', 'attached_at': '', 'detached_at': '', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'serial': '68c9b861-d548-4ed9-b618-463dfbd8cc16'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 959.768338] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ef49c88-bb56-4ac4-a30e-2c8623ec006f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.790501] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e7a5fc-94c9-48f8-b2fb-b0a834bdffa0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.798495] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a837df-769c-46d9-8341-8ed3988a23a0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.819013] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-212dd979-916a-4a84-88a6-1b9b552876c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.835149] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] The volume has not been displaced from its original location: [datastore1] volume-68c9b861-d548-4ed9-b618-463dfbd8cc16/volume-68c9b861-d548-4ed9-b618-463dfbd8cc16.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 959.840701] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Reconfiguring VM instance instance-0000005b to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 959.844786] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8dc6978d-96e7-4976-a8f9-97ee85aaaf06 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.865039] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 959.865039] env[61768]: value = "task-1229288" [ 959.865039] env[61768]: _type = "Task" [ 959.865039] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.875165] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229288, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.940932] env[61768]: DEBUG oslo_vmware.api [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229287, 'name': PowerOnVM_Task, 'duration_secs': 0.475051} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.944613] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 959.944916] env[61768]: INFO nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Took 9.08 seconds to spawn the instance on the hypervisor. [ 959.945238] env[61768]: DEBUG nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 959.947117] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c45215-0746-42dd-b39c-70574b847cc3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.959765] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b419ac-8393-4c76-83c4-efd9a52b532e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.968313] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb6e8ac-5544-4a2f-9cad-598d3e274360 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.001346] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "refresh_cache-271fe080-b8e9-4141-b2f3-4c4efaad81c7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.001550] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "refresh_cache-271fe080-b8e9-4141-b2f3-4c4efaad81c7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.001617] env[61768]: DEBUG nova.network.neutron [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 960.002865] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1419d1f7-520a-4a63-a248-875322375bbd tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.289s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.007020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04e13adc-e156-4a3c-8981-eafea5918580 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.013418] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752b96b0-9ad6-4c9a-9a04-d9514275bb97 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.028589] env[61768]: DEBUG nova.compute.provider_tree [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 960.136522] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524203f0-d2d4-38ba-f974-865ac61bcbf8, 'name': SearchDatastore_Task, 'duration_secs': 0.015132} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.136835] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.137094] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.137344] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.137501] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.137692] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.137951] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a77f1868-483d-4090-80b1-3bebbe53d5aa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.145980] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.146183] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 960.146862] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2ba9148-99a2-4dd7-acbd-992224d6c418 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.152094] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 960.152094] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a20203-49fd-07ea-2f91-707b5c42ce18" [ 960.152094] env[61768]: _type = "Task" [ 960.152094] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.161465] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a20203-49fd-07ea-2f91-707b5c42ce18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.379410] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229288, 'name': ReconfigVM_Task, 'duration_secs': 0.169889} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.379711] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Reconfigured VM instance instance-0000005b to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 960.384321] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3f965d2c-d6d6-4658-80ba-879f51e5763f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.400393] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 960.400393] env[61768]: value = "task-1229289" [ 960.400393] env[61768]: _type = "Task" [ 960.400393] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.408942] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229289, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.473688] env[61768]: INFO nova.compute.manager [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Took 16.62 seconds to build instance. [ 960.531500] env[61768]: DEBUG nova.scheduler.client.report [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.540661] env[61768]: DEBUG nova.network.neutron [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 960.662909] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a20203-49fd-07ea-2f91-707b5c42ce18, 'name': SearchDatastore_Task, 'duration_secs': 0.01132} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.663709] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ba1db66-a1c1-4bc4-b7e4-343552f6b82a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.668844] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 960.668844] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5263a7ec-fb18-f94b-5093-b50d0255df45" [ 960.668844] env[61768]: _type = "Task" [ 960.668844] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.676069] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5263a7ec-fb18-f94b-5093-b50d0255df45, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.734894] env[61768]: DEBUG nova.network.neutron [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Updating instance_info_cache with network_info: [{"id": "52cbc403-960a-4e8e-ab52-b1ea176a6de1", "address": "fa:16:3e:4d:d0:99", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52cbc403-96", "ovs_interfaceid": "52cbc403-960a-4e8e-ab52-b1ea176a6de1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.912939] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229289, 'name': ReconfigVM_Task, 'duration_secs': 0.210167} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.913631] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265498', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'name': 'volume-68c9b861-d548-4ed9-b618-463dfbd8cc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '30f55051-63b3-47a9-83fb-945cdeb82574', 'attached_at': '', 'detached_at': '', 'volume_id': '68c9b861-d548-4ed9-b618-463dfbd8cc16', 'serial': '68c9b861-d548-4ed9-b618-463dfbd8cc16'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 960.914043] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 960.915015] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4030a123-6a75-4a55-9d32-81efb69cc0b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.922028] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 960.922292] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fb2cf0cd-aa0a-4d0a-8967-512b46a9686f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.975652] env[61768]: DEBUG oslo_concurrency.lockutils [None req-361c7d5d-3ac1-449b-a6f4-65762ba02af3 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.137s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.978594] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 960.978829] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 960.979032] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Deleting the datastore file [datastore1] 30f55051-63b3-47a9-83fb-945cdeb82574 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.979305] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0a71302-abf7-42a0-ba99-c96d879a5444 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.987099] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for the task: (returnval){ [ 960.987099] env[61768]: value = "task-1229291" [ 960.987099] env[61768]: _type = "Task" [ 960.987099] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.995250] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229291, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.017327] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "8e0e13a7-eac5-4176-8536-2906d13c390e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.017546] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.017838] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "8e0e13a7-eac5-4176-8536-2906d13c390e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.018063] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.018259] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.020556] env[61768]: INFO nova.compute.manager [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Terminating instance [ 961.022475] env[61768]: DEBUG nova.compute.manager [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.022706] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 961.023633] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9e5398-ea3c-4485-8fcc-6654a0f8bd8f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.030886] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 961.031131] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-199c501d-6a04-404b-8200-fee503c79dc6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.036205] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.037731] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.337s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.038037] env[61768]: DEBUG nova.objects.instance [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'resources' on Instance uuid 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.039832] env[61768]: DEBUG oslo_vmware.api [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 961.039832] env[61768]: value = "task-1229292" [ 961.039832] env[61768]: _type = "Task" [ 961.039832] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.051111] env[61768]: DEBUG oslo_vmware.api [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229292, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.179119] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5263a7ec-fb18-f94b-5093-b50d0255df45, 'name': SearchDatastore_Task, 'duration_secs': 0.022484} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.179401] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.179674] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 961.179941] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-032a4da1-8f21-4f74-a532-ea25d54b738a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.186136] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 961.186136] env[61768]: value = "task-1229293" [ 961.186136] env[61768]: _type = "Task" [ 961.186136] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.194283] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229293, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.240221] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "refresh_cache-271fe080-b8e9-4141-b2f3-4c4efaad81c7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.240766] env[61768]: DEBUG nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Instance network_info: |[{"id": "52cbc403-960a-4e8e-ab52-b1ea176a6de1", "address": "fa:16:3e:4d:d0:99", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52cbc403-96", "ovs_interfaceid": "52cbc403-960a-4e8e-ab52-b1ea176a6de1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 961.241311] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4d:d0:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '52cbc403-960a-4e8e-ab52-b1ea176a6de1', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 961.249102] env[61768]: DEBUG oslo.service.loopingcall [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.249374] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 961.249614] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d8db05bb-2e23-42a7-8f4a-979cc17cb284 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.269765] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 961.269765] env[61768]: value = "task-1229294" [ 961.269765] env[61768]: _type = "Task" [ 961.269765] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.277239] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229294, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.400676] env[61768]: DEBUG nova.compute.manager [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Received event network-changed-52cbc403-960a-4e8e-ab52-b1ea176a6de1 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.401221] env[61768]: DEBUG nova.compute.manager [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Refreshing instance network info cache due to event network-changed-52cbc403-960a-4e8e-ab52-b1ea176a6de1. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 961.401618] env[61768]: DEBUG oslo_concurrency.lockutils [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] Acquiring lock "refresh_cache-271fe080-b8e9-4141-b2f3-4c4efaad81c7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.401904] env[61768]: DEBUG oslo_concurrency.lockutils [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] Acquired lock "refresh_cache-271fe080-b8e9-4141-b2f3-4c4efaad81c7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.402233] env[61768]: DEBUG nova.network.neutron [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Refreshing network info cache for port 52cbc403-960a-4e8e-ab52-b1ea176a6de1 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.498452] env[61768]: DEBUG oslo_vmware.api [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Task: {'id': task-1229291, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.12372} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.498765] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.498965] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 961.499169] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 961.499363] env[61768]: INFO nova.compute.manager [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Took 2.26 seconds to destroy the instance on the hypervisor. [ 961.499635] env[61768]: DEBUG oslo.service.loopingcall [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.499835] env[61768]: DEBUG nova.compute.manager [-] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.499921] env[61768]: DEBUG nova.network.neutron [-] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 961.540705] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "0cd97441-882d-4d7e-aafb-cc00d05d5929" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.540986] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "0cd97441-882d-4d7e-aafb-cc00d05d5929" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.549948] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "0cd97441-882d-4d7e-aafb-cc00d05d5929" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.009s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.550565] env[61768]: DEBUG nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 961.559441] env[61768]: DEBUG oslo_vmware.api [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229292, 'name': PowerOffVM_Task, 'duration_secs': 0.22911} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.559721] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 961.559898] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 961.560190] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-43b6a20d-504e-4da3-9a46-2726c83f95df {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.628382] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 961.628731] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 961.628924] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleting the datastore file [datastore1] 8e0e13a7-eac5-4176-8536-2906d13c390e {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 961.629215] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-86499bb3-e062-4aa8-958c-63e84e412634 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.641215] env[61768]: DEBUG oslo_vmware.api [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 961.641215] env[61768]: value = "task-1229296" [ 961.641215] env[61768]: _type = "Task" [ 961.641215] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.658143] env[61768]: DEBUG oslo_vmware.api [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.697966] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229293, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.713471] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a0013d-f167-4e71-8bdd-7b90e97d75e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.721364] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53657f1b-d908-412e-839b-a85adddda85a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.754343] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84bef278-4830-49da-883a-c20c454d23ca {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.762856] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4cf854-d1a3-42ac-a2de-72cca71c7ced {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.767861] env[61768]: DEBUG nova.compute.manager [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Received event network-changed-3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.767952] env[61768]: DEBUG nova.compute.manager [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Refreshing instance network info cache due to event network-changed-3ceeacd4-686d-4a24-bafc-328b8bd344ed. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 961.768636] env[61768]: DEBUG oslo_concurrency.lockutils [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] Acquiring lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.768636] env[61768]: DEBUG oslo_concurrency.lockutils [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] Acquired lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.768636] env[61768]: DEBUG nova.network.neutron [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Refreshing network info cache for port 3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 961.781662] env[61768]: DEBUG nova.compute.provider_tree [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.788253] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229294, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.056029] env[61768]: DEBUG nova.compute.utils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.058923] env[61768]: DEBUG nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 962.060023] env[61768]: DEBUG nova.network.neutron [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 962.126521] env[61768]: DEBUG nova.policy [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '07ce37acc83943f5868522eea44080a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '55c9d9b119924bec9fe21f1854d95cb2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 962.154153] env[61768]: DEBUG oslo_vmware.api [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.441243} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.154728] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 962.154997] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 962.155577] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 962.155934] env[61768]: INFO nova.compute.manager [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Took 1.13 seconds to destroy the instance on the hypervisor. [ 962.156663] env[61768]: DEBUG oslo.service.loopingcall [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.156933] env[61768]: DEBUG nova.compute.manager [-] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.157094] env[61768]: DEBUG nova.network.neutron [-] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 962.199981] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229293, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.556034} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.200713] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 962.201059] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 962.201423] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1abaf5a4-807a-4880-8237-c239b43b9a9d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.209476] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 962.209476] env[61768]: value = "task-1229297" [ 962.209476] env[61768]: _type = "Task" [ 962.209476] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.222051] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229297, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.224128] env[61768]: DEBUG nova.network.neutron [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Updated VIF entry in instance network info cache for port 52cbc403-960a-4e8e-ab52-b1ea176a6de1. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.224128] env[61768]: DEBUG nova.network.neutron [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Updating instance_info_cache with network_info: [{"id": "52cbc403-960a-4e8e-ab52-b1ea176a6de1", "address": "fa:16:3e:4d:d0:99", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap52cbc403-96", "ovs_interfaceid": "52cbc403-960a-4e8e-ab52-b1ea176a6de1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.285916] env[61768]: DEBUG nova.scheduler.client.report [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.289523] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229294, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.481888] env[61768]: DEBUG nova.network.neutron [-] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.566247] env[61768]: DEBUG nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 962.724413] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229297, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188442} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.724703] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 962.725524] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d126bd0-e06b-4c16-84f0-10b9958d8485 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.732020] env[61768]: DEBUG oslo_concurrency.lockutils [req-2ba7f18e-6565-493d-8f9d-f2f26200ea82 req-a2e9d879-df26-4b13-8be5-cdc0de907add service nova] Releasing lock "refresh_cache-271fe080-b8e9-4141-b2f3-4c4efaad81c7" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.748794] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 962.752126] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d803c963-18df-4500-93f8-6a3d0e13180d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.774034] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 962.774034] env[61768]: value = "task-1229298" [ 962.774034] env[61768]: _type = "Task" [ 962.774034] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.785745] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229298, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.789293] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229294, 'name': CreateVM_Task, 'duration_secs': 1.306242} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.789367] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 962.790420] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.790655] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.791120] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.792132] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.794185] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a57d169-968c-4c1a-8296-c6242b2a7586 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.796117] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.094s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.797994] env[61768]: INFO nova.compute.claims [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 962.807340] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 962.807340] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522040e0-f0b0-e84d-c090-ea5759de7bc3" [ 962.807340] env[61768]: _type = "Task" [ 962.807340] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.816765] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522040e0-f0b0-e84d-c090-ea5759de7bc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.820306] env[61768]: INFO nova.scheduler.client.report [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted allocations for instance 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98 [ 962.844810] env[61768]: DEBUG nova.network.neutron [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Successfully created port: 00b15ed3-61dc-432b-999a-710e050fb689 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 962.939756] env[61768]: DEBUG nova.network.neutron [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updated VIF entry in instance network info cache for port 3ceeacd4-686d-4a24-bafc-328b8bd344ed. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.940151] env[61768]: DEBUG nova.network.neutron [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updating instance_info_cache with network_info: [{"id": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "address": "fa:16:3e:fb:f6:d9", "network": {"id": "f6b2adff-fbc7-4119-9881-92e247f8579e", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1291801240-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.160", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f844c9d775474fb69d9ea04406cc1632", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ccc845e3-654b-43c6-acea-dde1084f0ad0", "external-id": "nsx-vlan-transportzone-344", "segmentation_id": 344, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3ceeacd4-68", "ovs_interfaceid": "3ceeacd4-686d-4a24-bafc-328b8bd344ed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.988254] env[61768]: INFO nova.compute.manager [-] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Took 1.49 seconds to deallocate network for instance. [ 963.165276] env[61768]: DEBUG nova.network.neutron [-] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.285356] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229298, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.317558] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522040e0-f0b0-e84d-c090-ea5759de7bc3, 'name': SearchDatastore_Task, 'duration_secs': 0.032519} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.318476] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.318722] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.318967] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.319610] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.319610] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.319893] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-01a9300d-71ed-4b48-861a-3b486c5262b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.328179] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.328371] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 963.329365] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9a800cb1-2846-4cae-9291-d8b2614dbeb7 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "71ce99fa-3d9d-4ff6-b06a-2e27d575ab98" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.281s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.330229] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c577ef72-5af6-4c1c-9d70-1822c09d0585 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.336218] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 963.336218] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ba2ad7-0826-2d8f-c9d6-eb68b98dd9d2" [ 963.336218] env[61768]: _type = "Task" [ 963.336218] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.344202] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ba2ad7-0826-2d8f-c9d6-eb68b98dd9d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.443607] env[61768]: DEBUG oslo_concurrency.lockutils [req-01806bcd-2585-41af-b78b-1d184a575902 req-a527909d-1c07-48a6-9bc4-49bb8517b9bb service nova] Releasing lock "refresh_cache-e1e90c98-5a65-473c-b508-b45fd93e31ad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.446301] env[61768]: DEBUG nova.compute.manager [req-453b39a5-fd12-4534-a564-475605fae6c0 req-fa591735-d8b3-4873-b66f-1580cba80687 service nova] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Received event network-vif-deleted-4cd7e358-0ed6-4baa-82c8-bc51b3e1c7e3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.533455] env[61768]: INFO nova.compute.manager [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Took 0.54 seconds to detach 1 volumes for instance. [ 963.535694] env[61768]: DEBUG nova.compute.manager [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Deleting volume: 68c9b861-d548-4ed9-b618-463dfbd8cc16 {{(pid=61768) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 963.577338] env[61768]: DEBUG nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 963.603172] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 963.603450] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 963.603618] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 963.603808] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 963.603966] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 963.604308] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 963.604551] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 963.604726] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 963.604906] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 963.605099] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 963.605273] env[61768]: DEBUG nova.virt.hardware [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 963.606418] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b8b4e7-4dd6-44bc-a1ca-4ae734bbf2d7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.614829] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd171cde-965c-441c-a4a6-72cd4b77f5eb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.667538] env[61768]: INFO nova.compute.manager [-] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Took 1.51 seconds to deallocate network for instance. [ 963.784032] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229298, 'name': ReconfigVM_Task, 'duration_secs': 0.917294} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.784297] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 963.784940] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef60f11c-48a9-4b87-96b5-4435321f6592 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.792620] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 963.792620] env[61768]: value = "task-1229300" [ 963.792620] env[61768]: _type = "Task" [ 963.792620] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.802315] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229300, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.845456] env[61768]: DEBUG nova.compute.manager [req-e4ec0a46-038b-405f-856e-6f66e0fa862c req-8cea0c22-d3f8-404a-8f30-e177770a2604 service nova] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Received event network-vif-deleted-e02bb27b-f94f-4f93-859a-5af2646daaaa {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.852766] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ba2ad7-0826-2d8f-c9d6-eb68b98dd9d2, 'name': SearchDatastore_Task, 'duration_secs': 0.010046} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.853851] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36e2e046-47d1-4ffd-9d0c-18af027a9b56 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.861818] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 963.861818] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52be5c55-d9b5-491e-5bcb-91dae641c4ba" [ 963.861818] env[61768]: _type = "Task" [ 963.861818] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.871620] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52be5c55-d9b5-491e-5bcb-91dae641c4ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.915957] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe7aae0-8eae-4b8f-85f1-faba659d14ee {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.923318] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3541ec8b-7da6-4c91-aba3-408b14b2065e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.953493] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30f933a6-cb8b-4343-b6a9-b667003f3152 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.961690] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363fb1b2-676f-4d26-9e47-201c96d23815 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.976952] env[61768]: DEBUG nova.compute.provider_tree [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.073812] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.175316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.302783] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229300, 'name': Rename_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.372031] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52be5c55-d9b5-491e-5bcb-91dae641c4ba, 'name': SearchDatastore_Task, 'duration_secs': 0.04006} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.372318] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.372593] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 271fe080-b8e9-4141-b2f3-4c4efaad81c7/271fe080-b8e9-4141-b2f3-4c4efaad81c7.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.372866] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-316a752e-2e55-485c-aea8-e31793600856 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.379308] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 964.379308] env[61768]: value = "task-1229301" [ 964.379308] env[61768]: _type = "Task" [ 964.379308] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.379781] env[61768]: DEBUG nova.network.neutron [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Successfully updated port: 00b15ed3-61dc-432b-999a-710e050fb689 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 964.390453] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229301, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.480056] env[61768]: DEBUG nova.scheduler.client.report [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.645526] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.645772] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.805184] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229300, 'name': Rename_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.886645] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "refresh_cache-49723fff-d1ce-476f-ac23-bd2472762d94" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 964.886810] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquired lock "refresh_cache-49723fff-d1ce-476f-ac23-bd2472762d94" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.886969] env[61768]: DEBUG nova.network.neutron [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 964.891133] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229301, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.986032] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.190s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.986628] env[61768]: DEBUG nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 964.989461] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.059s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.989658] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.989860] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 964.990206] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.917s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.990432] env[61768]: DEBUG nova.objects.instance [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lazy-loading 'resources' on Instance uuid 30f55051-63b3-47a9-83fb-945cdeb82574 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 964.992301] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b45ec8-33b0-4f31-94d6-c05cd8ea0af0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.002841] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d4ed855-2536-40b5-a662-b6d1bea0a0d8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.020087] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2498960-17f1-4d2a-9ad2-ae273f96a7b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.029303] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c199b2d8-50bf-4177-8882-02956c3ec564 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.064664] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179482MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 965.064953] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.148664] env[61768]: DEBUG nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 965.303137] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229300, 'name': Rename_Task, 'duration_secs': 1.099757} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.303471] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 965.303698] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-43ec65bb-8d1a-4af7-843d-5278074ad8ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.309842] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 965.309842] env[61768]: value = "task-1229302" [ 965.309842] env[61768]: _type = "Task" [ 965.309842] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.317186] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229302, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.391118] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229301, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.878612} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.391398] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 271fe080-b8e9-4141-b2f3-4c4efaad81c7/271fe080-b8e9-4141-b2f3-4c4efaad81c7.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 965.391731] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 965.393607] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cce793fd-870f-407b-8ff4-72b20f596132 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.400900] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 965.400900] env[61768]: value = "task-1229303" [ 965.400900] env[61768]: _type = "Task" [ 965.400900] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.408830] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229303, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.431804] env[61768]: DEBUG nova.network.neutron [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 965.472599] env[61768]: DEBUG nova.compute.manager [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Received event network-vif-plugged-00b15ed3-61dc-432b-999a-710e050fb689 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.472918] env[61768]: DEBUG oslo_concurrency.lockutils [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] Acquiring lock "49723fff-d1ce-476f-ac23-bd2472762d94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.473146] env[61768]: DEBUG oslo_concurrency.lockutils [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] Lock "49723fff-d1ce-476f-ac23-bd2472762d94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.473411] env[61768]: DEBUG oslo_concurrency.lockutils [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] Lock "49723fff-d1ce-476f-ac23-bd2472762d94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.473625] env[61768]: DEBUG nova.compute.manager [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] No waiting events found dispatching network-vif-plugged-00b15ed3-61dc-432b-999a-710e050fb689 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 965.473809] env[61768]: WARNING nova.compute.manager [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Received unexpected event network-vif-plugged-00b15ed3-61dc-432b-999a-710e050fb689 for instance with vm_state building and task_state spawning. [ 965.473979] env[61768]: DEBUG nova.compute.manager [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Received event network-changed-00b15ed3-61dc-432b-999a-710e050fb689 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 965.474182] env[61768]: DEBUG nova.compute.manager [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Refreshing instance network info cache due to event network-changed-00b15ed3-61dc-432b-999a-710e050fb689. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 965.474365] env[61768]: DEBUG oslo_concurrency.lockutils [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] Acquiring lock "refresh_cache-49723fff-d1ce-476f-ac23-bd2472762d94" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.493179] env[61768]: DEBUG nova.compute.utils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 965.494483] env[61768]: DEBUG nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 965.494652] env[61768]: DEBUG nova.network.neutron [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 965.563584] env[61768]: DEBUG nova.policy [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cccca7ae3e764b7ead0419b0a4c24d04', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5499a3a9a9646f081c55e9ae3f9701e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 965.618763] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e125c835-ccdb-4001-832c-0f865d30ab2e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.631118] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23953fb-c8c0-488d-9dfd-ce0a5e1f640c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.677881] env[61768]: DEBUG nova.network.neutron [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Updating instance_info_cache with network_info: [{"id": "00b15ed3-61dc-432b-999a-710e050fb689", "address": "fa:16:3e:8c:3a:b8", "network": {"id": "6db24d01-da2d-4a6a-aeaa-552b38b68165", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1537675887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55c9d9b119924bec9fe21f1854d95cb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00b15ed3-61", "ovs_interfaceid": "00b15ed3-61dc-432b-999a-710e050fb689", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.679648] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7ae462-5f4c-453c-a509-0b9412a3aedb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.688115] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ffabad-4929-431a-b385-138a944a1e90 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.693463] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.702149] env[61768]: DEBUG nova.compute.provider_tree [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.820572] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229302, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.888453] env[61768]: DEBUG nova.network.neutron [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Successfully created port: 2be36d33-ba56-45fc-b047-be0c646b844d {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 965.910361] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229303, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077943} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.910707] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.911481] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-983f9322-804b-40ce-8b4f-5ddf207bf85a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.933041] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 271fe080-b8e9-4141-b2f3-4c4efaad81c7/271fe080-b8e9-4141-b2f3-4c4efaad81c7.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.933360] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afc95122-54b3-4538-b62a-4f27e27210f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.965327] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 965.965327] env[61768]: value = "task-1229304" [ 965.965327] env[61768]: _type = "Task" [ 965.965327] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.976040] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229304, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.997449] env[61768]: DEBUG nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 966.183691] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Releasing lock "refresh_cache-49723fff-d1ce-476f-ac23-bd2472762d94" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.184087] env[61768]: DEBUG nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Instance network_info: |[{"id": "00b15ed3-61dc-432b-999a-710e050fb689", "address": "fa:16:3e:8c:3a:b8", "network": {"id": "6db24d01-da2d-4a6a-aeaa-552b38b68165", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1537675887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55c9d9b119924bec9fe21f1854d95cb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00b15ed3-61", "ovs_interfaceid": "00b15ed3-61dc-432b-999a-710e050fb689", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 966.184463] env[61768]: DEBUG oslo_concurrency.lockutils [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] Acquired lock "refresh_cache-49723fff-d1ce-476f-ac23-bd2472762d94" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.184655] env[61768]: DEBUG nova.network.neutron [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Refreshing network info cache for port 00b15ed3-61dc-432b-999a-710e050fb689 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 966.185947] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8c:3a:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00b15ed3-61dc-432b-999a-710e050fb689', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.195096] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Creating folder: Project (55c9d9b119924bec9fe21f1854d95cb2). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 966.196175] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6b21ea4f-c136-4d64-8131-dedc564ee472 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.204539] env[61768]: DEBUG nova.scheduler.client.report [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.209018] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Created folder: Project (55c9d9b119924bec9fe21f1854d95cb2) in parent group-v265360. [ 966.209233] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Creating folder: Instances. Parent ref: group-v265519. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 966.209691] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-626e94f0-64d2-4ab4-99c1-b63811cf236f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.220368] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Created folder: Instances in parent group-v265519. [ 966.220653] env[61768]: DEBUG oslo.service.loopingcall [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.221427] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 966.221693] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45bc3bcf-a848-492a-ac6a-f37ee9dfa8bb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.243022] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.243022] env[61768]: value = "task-1229307" [ 966.243022] env[61768]: _type = "Task" [ 966.243022] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.252125] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229307, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.324234] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229302, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.475533] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229304, 'name': ReconfigVM_Task, 'duration_secs': 0.292829} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.475823] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 271fe080-b8e9-4141-b2f3-4c4efaad81c7/271fe080-b8e9-4141-b2f3-4c4efaad81c7.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 966.476523] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3e038384-4061-4dca-a9a4-2c00c022a223 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.483707] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 966.483707] env[61768]: value = "task-1229308" [ 966.483707] env[61768]: _type = "Task" [ 966.483707] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.491827] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229308, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.710880] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.713066] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.538s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.713324] env[61768]: DEBUG nova.objects.instance [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'resources' on Instance uuid 8e0e13a7-eac5-4176-8536-2906d13c390e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 966.738638] env[61768]: INFO nova.scheduler.client.report [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Deleted allocations for instance 30f55051-63b3-47a9-83fb-945cdeb82574 [ 966.756902] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229307, 'name': CreateVM_Task, 'duration_secs': 0.412358} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.758918] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 966.758918] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.758918] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.758918] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 966.758918] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3aa6d82-9a9b-4577-9cb5-f0f1231acce4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.763265] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 966.763265] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526e2db6-c5fc-ec67-e70c-a155b902e2ab" [ 966.763265] env[61768]: _type = "Task" [ 966.763265] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.771025] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526e2db6-c5fc-ec67-e70c-a155b902e2ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.820836] env[61768]: DEBUG oslo_vmware.api [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229302, 'name': PowerOnVM_Task, 'duration_secs': 1.051331} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.821160] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 966.821382] env[61768]: INFO nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Took 11.90 seconds to spawn the instance on the hypervisor. [ 966.821573] env[61768]: DEBUG nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.822590] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5379a7a7-ac75-441d-9b60-254c08c1c3f5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.962903] env[61768]: DEBUG nova.network.neutron [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Updated VIF entry in instance network info cache for port 00b15ed3-61dc-432b-999a-710e050fb689. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 966.963307] env[61768]: DEBUG nova.network.neutron [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Updating instance_info_cache with network_info: [{"id": "00b15ed3-61dc-432b-999a-710e050fb689", "address": "fa:16:3e:8c:3a:b8", "network": {"id": "6db24d01-da2d-4a6a-aeaa-552b38b68165", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1537675887-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "55c9d9b119924bec9fe21f1854d95cb2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bde2f6cc-fb26-4d71-95a6-57d1ae1c4afd", "external-id": "nsx-vlan-transportzone-206", "segmentation_id": 206, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00b15ed3-61", "ovs_interfaceid": "00b15ed3-61dc-432b-999a-710e050fb689", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.995290] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229308, 'name': Rename_Task, 'duration_secs': 0.134408} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.995565] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 966.995816] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-065620a4-3448-47cf-a62c-0877cff9fabe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.002456] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 967.002456] env[61768]: value = "task-1229309" [ 967.002456] env[61768]: _type = "Task" [ 967.002456] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.006377] env[61768]: DEBUG nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 967.011495] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.033449] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.033706] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.033905] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.034095] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.034256] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.034416] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.034631] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.034820] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.035013] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.035194] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.035380] env[61768]: DEBUG nova.virt.hardware [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.036288] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4483345-c953-4863-b0aa-f190dbca6dda {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.044336] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefad1bd-c9f9-4e18-a981-60dd874f1ff5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.250041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-73b62d3e-0fdf-4195-bb29-50fcc514e6ad tempest-ServersTestBootFromVolume-1572134970 tempest-ServersTestBootFromVolume-1572134970-project-member] Lock "30f55051-63b3-47a9-83fb-945cdeb82574" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.016s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.277627] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]526e2db6-c5fc-ec67-e70c-a155b902e2ab, 'name': SearchDatastore_Task, 'duration_secs': 0.012428} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.278152] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.278400] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 967.278660] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.278826] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.279041] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 967.279284] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fb36e538-16c1-433c-94d3-2d5e92acd4bc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.289646] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 967.289843] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 967.290607] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05bf8466-1812-4436-ae86-d806fda49f3b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.296429] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 967.296429] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52535cec-eee6-e89f-90c1-7e9102dd7798" [ 967.296429] env[61768]: _type = "Task" [ 967.296429] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.306392] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52535cec-eee6-e89f-90c1-7e9102dd7798, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.341675] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438d0c6b-0cc3-435f-bf88-c3bef60366c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.343651] env[61768]: INFO nova.compute.manager [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Took 19.35 seconds to build instance. [ 967.349055] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b83b3c5-9e20-46a1-bd05-947ac75d28c5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.381435] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f310ff0-6569-4062-a870-f7980d7728b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.388937] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73fc9751-4a87-4599-9c2a-24abe794dd9a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.394966] env[61768]: DEBUG nova.compute.manager [req-ec83f808-22e6-4f85-909a-fcba2df069dd req-301295de-cdf0-4e20-a5ab-def3ef8fd82f service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Received event network-vif-plugged-2be36d33-ba56-45fc-b047-be0c646b844d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.395203] env[61768]: DEBUG oslo_concurrency.lockutils [req-ec83f808-22e6-4f85-909a-fcba2df069dd req-301295de-cdf0-4e20-a5ab-def3ef8fd82f service nova] Acquiring lock "750bc93c-65aa-4afd-a07b-cbeda0acae24-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.395416] env[61768]: DEBUG oslo_concurrency.lockutils [req-ec83f808-22e6-4f85-909a-fcba2df069dd req-301295de-cdf0-4e20-a5ab-def3ef8fd82f service nova] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.395591] env[61768]: DEBUG oslo_concurrency.lockutils [req-ec83f808-22e6-4f85-909a-fcba2df069dd req-301295de-cdf0-4e20-a5ab-def3ef8fd82f service nova] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.395768] env[61768]: DEBUG nova.compute.manager [req-ec83f808-22e6-4f85-909a-fcba2df069dd req-301295de-cdf0-4e20-a5ab-def3ef8fd82f service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] No waiting events found dispatching network-vif-plugged-2be36d33-ba56-45fc-b047-be0c646b844d {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 967.395938] env[61768]: WARNING nova.compute.manager [req-ec83f808-22e6-4f85-909a-fcba2df069dd req-301295de-cdf0-4e20-a5ab-def3ef8fd82f service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Received unexpected event network-vif-plugged-2be36d33-ba56-45fc-b047-be0c646b844d for instance with vm_state building and task_state spawning. [ 967.406371] env[61768]: DEBUG nova.compute.provider_tree [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.466796] env[61768]: DEBUG oslo_concurrency.lockutils [req-650d7291-c801-4068-800f-37a803c326c4 req-927afa9b-82e4-4251-8a95-5d9d588cbcae service nova] Releasing lock "refresh_cache-49723fff-d1ce-476f-ac23-bd2472762d94" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.513056] env[61768]: DEBUG oslo_vmware.api [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229309, 'name': PowerOnVM_Task, 'duration_secs': 0.473939} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.514085] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 967.514085] env[61768]: INFO nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Took 8.46 seconds to spawn the instance on the hypervisor. [ 967.514085] env[61768]: DEBUG nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.514494] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39d99f2-dd27-46ab-98e7-15443acb9750 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.809034] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52535cec-eee6-e89f-90c1-7e9102dd7798, 'name': SearchDatastore_Task, 'duration_secs': 0.009127} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.809034] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96678ebb-7e8c-4ebc-aebb-e356ccd8a405 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.814548] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 967.814548] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52037e39-b4c4-9adc-7985-dcdfca2ff030" [ 967.814548] env[61768]: _type = "Task" [ 967.814548] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.823375] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52037e39-b4c4-9adc-7985-dcdfca2ff030, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.846232] env[61768]: DEBUG oslo_concurrency.lockutils [None req-6c794747-d2f8-46a0-9a46-e78f0134c98c tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.868s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.909870] env[61768]: DEBUG nova.scheduler.client.report [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.976528] env[61768]: DEBUG nova.network.neutron [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Successfully updated port: 2be36d33-ba56-45fc-b047-be0c646b844d {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.033201] env[61768]: INFO nova.compute.manager [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Took 18.40 seconds to build instance. [ 968.086782] env[61768]: DEBUG nova.compute.manager [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Received event network-changed-2be36d33-ba56-45fc-b047-be0c646b844d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.087036] env[61768]: DEBUG nova.compute.manager [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Refreshing instance network info cache due to event network-changed-2be36d33-ba56-45fc-b047-be0c646b844d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 968.087284] env[61768]: DEBUG oslo_concurrency.lockutils [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] Acquiring lock "refresh_cache-750bc93c-65aa-4afd-a07b-cbeda0acae24" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.087439] env[61768]: DEBUG oslo_concurrency.lockutils [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] Acquired lock "refresh_cache-750bc93c-65aa-4afd-a07b-cbeda0acae24" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.087608] env[61768]: DEBUG nova.network.neutron [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Refreshing network info cache for port 2be36d33-ba56-45fc-b047-be0c646b844d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.328093] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52037e39-b4c4-9adc-7985-dcdfca2ff030, 'name': SearchDatastore_Task, 'duration_secs': 0.021176} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.328379] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.328688] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 49723fff-d1ce-476f-ac23-bd2472762d94/49723fff-d1ce-476f-ac23-bd2472762d94.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 968.328959] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ab90f9f-c3fa-44e3-a1ef-11003e39e9a5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.335698] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 968.335698] env[61768]: value = "task-1229310" [ 968.335698] env[61768]: _type = "Task" [ 968.335698] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.343836] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229310, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.417020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.704s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.419618] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.355s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.446800] env[61768]: INFO nova.scheduler.client.report [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted allocations for instance 8e0e13a7-eac5-4176-8536-2906d13c390e [ 968.480548] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "refresh_cache-750bc93c-65aa-4afd-a07b-cbeda0acae24" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.534515] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9341ea45-b884-41b3-8dbd-d76b4e8ec217 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.909s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.624069] env[61768]: DEBUG nova.compute.manager [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.624359] env[61768]: DEBUG nova.compute.manager [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing instance network info cache due to event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 968.624525] env[61768]: DEBUG oslo_concurrency.lockutils [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.624681] env[61768]: DEBUG oslo_concurrency.lockutils [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.624852] env[61768]: DEBUG nova.network.neutron [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing network info cache for port 8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 968.652279] env[61768]: DEBUG nova.network.neutron [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 968.828720] env[61768]: DEBUG nova.network.neutron [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 968.850757] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229310, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.956663] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c78658e8-8598-46df-8460-b558277ee454 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "8e0e13a7-eac5-4176-8536-2906d13c390e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.939s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.331274] env[61768]: DEBUG oslo_concurrency.lockutils [req-e6d0da86-e5dc-41c5-a5b5-ac7e90799ccc req-af7aff4f-bef6-4cd7-9efb-2449c459b2d7 service nova] Releasing lock "refresh_cache-750bc93c-65aa-4afd-a07b-cbeda0acae24" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.331704] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquired lock "refresh_cache-750bc93c-65aa-4afd-a07b-cbeda0acae24" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.331880] env[61768]: DEBUG nova.network.neutron [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 969.352591] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229310, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.527296} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.353582] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 49723fff-d1ce-476f-ac23-bd2472762d94/49723fff-d1ce-476f-ac23-bd2472762d94.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 969.354302] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 969.354302] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c976942d-44e1-4d4d-9379-5e2c0df59b88 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.366413] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 969.366413] env[61768]: value = "task-1229311" [ 969.366413] env[61768]: _type = "Task" [ 969.366413] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.381396] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229311, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.450815] env[61768]: DEBUG nova.network.neutron [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updated VIF entry in instance network info cache for port 8e676cc0-ee67-4e76-afe1-2dff2e254617. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 969.450815] env[61768]: DEBUG nova.network.neutron [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.461677] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 2c85de5d-0111-40a6-a8bc-69c7eba0393d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.461677] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance e1e90c98-5a65-473c-b508-b45fd93e31ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.461677] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 81641ce1-cc88-4df1-a54c-48f9bd4bb73b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.461677] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 271fe080-b8e9-4141-b2f3-4c4efaad81c7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.461677] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 49723fff-d1ce-476f-ac23-bd2472762d94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.461677] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 750bc93c-65aa-4afd-a07b-cbeda0acae24 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 969.511491] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.511491] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.511491] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.511491] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.511750] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.514216] env[61768]: INFO nova.compute.manager [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Terminating instance [ 969.516286] env[61768]: DEBUG nova.compute.manager [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 969.516286] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 969.517101] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cca3a45e-dc10-48c5-b22b-3ed0647697e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.525480] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 969.525756] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e35b181-5f95-4779-8771-3ef14623ac62 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.533070] env[61768]: DEBUG oslo_vmware.api [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 969.533070] env[61768]: value = "task-1229312" [ 969.533070] env[61768]: _type = "Task" [ 969.533070] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.540970] env[61768]: DEBUG oslo_vmware.api [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229312, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.876857] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229311, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065483} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.877397] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 969.878322] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50eeddd1-6c79-421e-b578-62b7279eccb8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.905016] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 49723fff-d1ce-476f-ac23-bd2472762d94/49723fff-d1ce-476f-ac23-bd2472762d94.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 969.905016] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a98fd20f-9171-4897-8750-b95ce9d1dcb8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.922018] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 969.922018] env[61768]: value = "task-1229313" [ 969.922018] env[61768]: _type = "Task" [ 969.922018] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.929914] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229313, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.953723] env[61768]: DEBUG nova.network.neutron [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 969.954227] env[61768]: DEBUG oslo_concurrency.lockutils [req-2d01c9fd-a78c-4b6b-9c00-56630522ce4e req-cf592b60-fd9e-46bb-a45e-67c7b9087017 service nova] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.965588] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 969.965817] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 969.965971] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 970.046081] env[61768]: DEBUG oslo_vmware.api [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229312, 'name': PowerOffVM_Task, 'duration_secs': 0.204863} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.046503] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 970.046801] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 970.047557] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4752f653-c3cd-4d6e-8d86-04485c2a3882 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.083358] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394d7a93-ea61-4cd6-9363-c84a6c59d47a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.094024] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40eae02-f37b-4aef-8df4-94d524057b76 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.124756] env[61768]: DEBUG nova.network.neutron [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Updating instance_info_cache with network_info: [{"id": "2be36d33-ba56-45fc-b047-be0c646b844d", "address": "fa:16:3e:5c:52:e3", "network": {"id": "e7cf36c3-7f67-4615-901d-09e669470743", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1144343643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5499a3a9a9646f081c55e9ae3f9701e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2be36d33-ba", "ovs_interfaceid": "2be36d33-ba56-45fc-b047-be0c646b844d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.126493] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedd7b0f-2aa2-415e-bf7b-45cdb17c80f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.134189] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4492c89-b3f9-4583-89a4-d45fb60c4f19 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.148111] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.222963] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 970.223269] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 970.223461] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleting the datastore file [datastore2] 271fe080-b8e9-4141-b2f3-4c4efaad81c7 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 970.223732] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7f5e1e9f-527a-4153-86a7-569528cdc687 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.230586] env[61768]: DEBUG oslo_vmware.api [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 970.230586] env[61768]: value = "task-1229315" [ 970.230586] env[61768]: _type = "Task" [ 970.230586] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.238522] env[61768]: DEBUG oslo_vmware.api [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229315, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.432679] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229313, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.629803] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Releasing lock "refresh_cache-750bc93c-65aa-4afd-a07b-cbeda0acae24" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.630166] env[61768]: DEBUG nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Instance network_info: |[{"id": "2be36d33-ba56-45fc-b047-be0c646b844d", "address": "fa:16:3e:5c:52:e3", "network": {"id": "e7cf36c3-7f67-4615-901d-09e669470743", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1144343643-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f5499a3a9a9646f081c55e9ae3f9701e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d1da5fc2-0280-4f76-ac97-20ea4bc7bb16", "external-id": "nsx-vlan-transportzone-563", "segmentation_id": 563, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2be36d33-ba", "ovs_interfaceid": "2be36d33-ba56-45fc-b047-be0c646b844d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.630682] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5c:52:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd1da5fc2-0280-4f76-ac97-20ea4bc7bb16', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2be36d33-ba56-45fc-b047-be0c646b844d', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.638087] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Creating folder: Project (f5499a3a9a9646f081c55e9ae3f9701e). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 970.638371] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fc527523-fa76-496b-8811-50ef975d5c0f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.648664] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Created folder: Project (f5499a3a9a9646f081c55e9ae3f9701e) in parent group-v265360. [ 970.648851] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Creating folder: Instances. Parent ref: group-v265522. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 970.649091] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-031b655b-26cb-4e6f-b73c-51730dc8a018 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.651189] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.661505] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Created folder: Instances in parent group-v265522. [ 970.661729] env[61768]: DEBUG oslo.service.loopingcall [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.661966] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 970.662195] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08030fa2-3366-46f2-9619-3d96e681ac77 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.682071] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.682071] env[61768]: value = "task-1229318" [ 970.682071] env[61768]: _type = "Task" [ 970.682071] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.689672] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229318, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.743497] env[61768]: DEBUG oslo_vmware.api [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229315, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228436} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.743780] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 970.743977] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 970.744179] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 970.744391] env[61768]: INFO nova.compute.manager [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Took 1.23 seconds to destroy the instance on the hypervisor. [ 970.744674] env[61768]: DEBUG oslo.service.loopingcall [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.744845] env[61768]: DEBUG nova.compute.manager [-] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.744940] env[61768]: DEBUG nova.network.neutron [-] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 970.933939] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229313, 'name': ReconfigVM_Task, 'duration_secs': 0.618792} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.937580] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 49723fff-d1ce-476f-ac23-bd2472762d94/49723fff-d1ce-476f-ac23-bd2472762d94.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.938017] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-07b13e65-335b-4885-aeed-43ceaa3f3dd0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.948879] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 970.948879] env[61768]: value = "task-1229319" [ 970.948879] env[61768]: _type = "Task" [ 970.948879] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.964843] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229319, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.975248] env[61768]: DEBUG nova.compute.manager [req-63e7c6ff-dc23-4cd2-a5f3-90316881dc4c req-6ad6a698-9b6d-4ff0-9c56-dd896cb4a791 service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Received event network-vif-deleted-52cbc403-960a-4e8e-ab52-b1ea176a6de1 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.975558] env[61768]: INFO nova.compute.manager [req-63e7c6ff-dc23-4cd2-a5f3-90316881dc4c req-6ad6a698-9b6d-4ff0-9c56-dd896cb4a791 service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Neutron deleted interface 52cbc403-960a-4e8e-ab52-b1ea176a6de1; detaching it from the instance and deleting it from the info cache [ 970.975672] env[61768]: DEBUG nova.network.neutron [req-63e7c6ff-dc23-4cd2-a5f3-90316881dc4c req-6ad6a698-9b6d-4ff0-9c56-dd896cb4a791 service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.156378] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 971.156612] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.737s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.156911] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.464s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.158412] env[61768]: INFO nova.compute.claims [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.192439] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229318, 'name': CreateVM_Task, 'duration_secs': 0.315723} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.192617] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 971.193328] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.193507] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.193842] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.194122] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79822961-de62-48b7-89a3-24114b262dbc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.198713] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 971.198713] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a3254-069c-e7a3-7fa2-cdfd2c899c8b" [ 971.198713] env[61768]: _type = "Task" [ 971.198713] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.206695] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a3254-069c-e7a3-7fa2-cdfd2c899c8b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.455547] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229319, 'name': Rename_Task, 'duration_secs': 0.190014} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.455832] env[61768]: DEBUG nova.network.neutron [-] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.457064] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 971.457362] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-673c583f-e094-4d40-90b5-f08997b312ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.464456] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 971.464456] env[61768]: value = "task-1229320" [ 971.464456] env[61768]: _type = "Task" [ 971.464456] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.472401] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229320, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.478357] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3a8331af-21dc-4295-b741-d391c8615882 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.486709] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e9eff3-1c48-46d1-bcab-60f2ef913073 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.512315] env[61768]: DEBUG nova.compute.manager [req-63e7c6ff-dc23-4cd2-a5f3-90316881dc4c req-6ad6a698-9b6d-4ff0-9c56-dd896cb4a791 service nova] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Detach interface failed, port_id=52cbc403-960a-4e8e-ab52-b1ea176a6de1, reason: Instance 271fe080-b8e9-4141-b2f3-4c4efaad81c7 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 971.667935] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "86f41ef0-4c92-4626-bb83-1148144c597c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.668564] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.709185] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a3254-069c-e7a3-7fa2-cdfd2c899c8b, 'name': SearchDatastore_Task, 'duration_secs': 0.025797} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.709493] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.709729] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.709968] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.710144] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.710334] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.710831] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-deaaf5ca-3555-42f6-a452-adad7d184a6f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.720942] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.721146] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 971.721833] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cd05ed6-5425-4560-8a1c-01191d04b96f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.726451] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 971.726451] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527e1f66-f6fc-c742-72bb-2fdd5c29f5e4" [ 971.726451] env[61768]: _type = "Task" [ 971.726451] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.733514] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527e1f66-f6fc-c742-72bb-2fdd5c29f5e4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.960749] env[61768]: INFO nova.compute.manager [-] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Took 1.22 seconds to deallocate network for instance. [ 971.974723] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229320, 'name': PowerOnVM_Task} progress is 76%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.170125] env[61768]: DEBUG nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.236632] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527e1f66-f6fc-c742-72bb-2fdd5c29f5e4, 'name': SearchDatastore_Task, 'duration_secs': 0.010897} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.237454] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cae676d5-cd82-4838-91b7-4aca5ac11d97 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.243995] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 972.243995] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c1dfba-4567-71b8-d223-2bb397699013" [ 972.243995] env[61768]: _type = "Task" [ 972.243995] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.251826] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c1dfba-4567-71b8-d223-2bb397699013, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.276921] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b95ec8b-f9bc-4698-936c-61838b5f6c64 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.283987] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d4ec64-64b2-40b2-aa05-9df87bfd40b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.314095] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff277b4d-d90f-4f19-b7b9-eee46287423c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.321345] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702bf5d7-5b3b-447b-9c9e-14706acdc135 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.334548] env[61768]: DEBUG nova.compute.provider_tree [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.467136] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.476350] env[61768]: DEBUG oslo_vmware.api [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229320, 'name': PowerOnVM_Task, 'duration_secs': 0.907331} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.476616] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 972.476793] env[61768]: INFO nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Took 8.90 seconds to spawn the instance on the hypervisor. [ 972.477338] env[61768]: DEBUG nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.477933] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e634b90-5cb8-4ee3-af99-7bdd2128af5b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.689118] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.754916] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c1dfba-4567-71b8-d223-2bb397699013, 'name': SearchDatastore_Task, 'duration_secs': 0.013958} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.755221] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.755492] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 750bc93c-65aa-4afd-a07b-cbeda0acae24/750bc93c-65aa-4afd-a07b-cbeda0acae24.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 972.755756] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-30c21181-4e89-4d4f-a54e-0056ae4c9b7d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.762821] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 972.762821] env[61768]: value = "task-1229321" [ 972.762821] env[61768]: _type = "Task" [ 972.762821] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.770804] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229321, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.838657] env[61768]: DEBUG nova.scheduler.client.report [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.996170] env[61768]: INFO nova.compute.manager [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Took 21.15 seconds to build instance. [ 973.272969] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229321, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.439063} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.273242] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 750bc93c-65aa-4afd-a07b-cbeda0acae24/750bc93c-65aa-4afd-a07b-cbeda0acae24.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 973.273465] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.273717] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-78610aa9-b34d-4820-bc1e-92ad58d0f1c3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.280082] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 973.280082] env[61768]: value = "task-1229322" [ 973.280082] env[61768]: _type = "Task" [ 973.280082] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.288043] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229322, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.315614] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "49723fff-d1ce-476f-ac23-bd2472762d94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.343158] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.186s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.343696] env[61768]: DEBUG nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.346334] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.879s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.346617] env[61768]: DEBUG nova.objects.instance [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'resources' on Instance uuid 271fe080-b8e9-4141-b2f3-4c4efaad81c7 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.498571] env[61768]: DEBUG oslo_concurrency.lockutils [None req-bc9abeb1-ed39-4211-a6f9-3ec7573f6b9a tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "49723fff-d1ce-476f-ac23-bd2472762d94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.683s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.498969] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "49723fff-d1ce-476f-ac23-bd2472762d94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.183s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.499280] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "49723fff-d1ce-476f-ac23-bd2472762d94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.499556] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "49723fff-d1ce-476f-ac23-bd2472762d94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.499786] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "49723fff-d1ce-476f-ac23-bd2472762d94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.502145] env[61768]: INFO nova.compute.manager [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Terminating instance [ 973.504089] env[61768]: DEBUG nova.compute.manager [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 973.504342] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 973.505261] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-912ac963-3f1c-4d60-952d-1d72a4796b9c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.513313] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 973.513582] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4bf326cb-be20-47f0-8c15-81b5e5684297 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.519906] env[61768]: DEBUG oslo_vmware.api [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 973.519906] env[61768]: value = "task-1229323" [ 973.519906] env[61768]: _type = "Task" [ 973.519906] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.527336] env[61768]: DEBUG oslo_vmware.api [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.789643] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229322, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064897} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.789865] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 973.790625] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d67bdc-e29a-4a04-8f94-04c269713c46 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.814964] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] 750bc93c-65aa-4afd-a07b-cbeda0acae24/750bc93c-65aa-4afd-a07b-cbeda0acae24.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.815290] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7ccc78f3-637c-4891-80a1-3ca67266769b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.836201] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 973.836201] env[61768]: value = "task-1229324" [ 973.836201] env[61768]: _type = "Task" [ 973.836201] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.845573] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229324, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.849284] env[61768]: DEBUG nova.compute.utils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.853594] env[61768]: DEBUG nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.853790] env[61768]: DEBUG nova.network.neutron [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 973.893990] env[61768]: DEBUG nova.policy [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd1aa7c174648888b58f6a6df6f82a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfce0cf4a5d14042a929dac65aeb2ae3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.968764] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1613dd20-b133-48cd-a189-08c8a4fa4f0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.976713] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d97cb07-480c-498a-937a-b6f1c6316d02 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.007608] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d303b76a-a1fe-47d0-9bc2-bce804b45556 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.015333] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d066ddb-b488-4d36-834f-2be948ab1ecb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.033857] env[61768]: DEBUG nova.compute.provider_tree [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 974.037881] env[61768]: DEBUG oslo_vmware.api [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229323, 'name': PowerOffVM_Task, 'duration_secs': 0.354188} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.038415] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 974.038612] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 974.038881] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8d58ddc-0521-4b80-aaf1-0d9e08fbe11d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.121084] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 974.121311] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 974.121505] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Deleting the datastore file [datastore2] 49723fff-d1ce-476f-ac23-bd2472762d94 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.121791] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8fa13d2e-b416-4c25-b8c3-b62a9390fd79 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.129484] env[61768]: DEBUG oslo_vmware.api [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for the task: (returnval){ [ 974.129484] env[61768]: value = "task-1229326" [ 974.129484] env[61768]: _type = "Task" [ 974.129484] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.137113] env[61768]: DEBUG oslo_vmware.api [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229326, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.213876] env[61768]: DEBUG nova.network.neutron [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Successfully created port: a845dca4-da7a-491b-a5a8-c83fb17f582b {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.346801] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229324, 'name': ReconfigVM_Task, 'duration_secs': 0.303506} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.347193] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Reconfigured VM instance instance-00000065 to attach disk [datastore1] 750bc93c-65aa-4afd-a07b-cbeda0acae24/750bc93c-65aa-4afd-a07b-cbeda0acae24.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.347872] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-976ae145-d316-4df3-9c90-e4ea1a80da1b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.354452] env[61768]: DEBUG nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.357035] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 974.357035] env[61768]: value = "task-1229327" [ 974.357035] env[61768]: _type = "Task" [ 974.357035] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.367318] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229327, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.540061] env[61768]: DEBUG nova.scheduler.client.report [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.638611] env[61768]: DEBUG oslo_vmware.api [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Task: {'id': task-1229326, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.34557} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.638850] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.639061] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 974.639253] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 974.639443] env[61768]: INFO nova.compute.manager [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Took 1.14 seconds to destroy the instance on the hypervisor. [ 974.639685] env[61768]: DEBUG oslo.service.loopingcall [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.639882] env[61768]: DEBUG nova.compute.manager [-] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 974.639978] env[61768]: DEBUG nova.network.neutron [-] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 974.871380] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229327, 'name': Rename_Task, 'duration_secs': 0.170694} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.871748] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 974.871983] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d2a4f41-e81a-42fc-81d0-2fe1b97c5963 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.878940] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 974.878940] env[61768]: value = "task-1229328" [ 974.878940] env[61768]: _type = "Task" [ 974.878940] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.886819] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229328, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.952834] env[61768]: DEBUG nova.compute.manager [req-77c37dad-a790-4219-ab58-39ba2e54572f req-ee6333ca-178e-4136-8bbf-c95822c64e95 service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Received event network-vif-deleted-00b15ed3-61dc-432b-999a-710e050fb689 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 974.953069] env[61768]: INFO nova.compute.manager [req-77c37dad-a790-4219-ab58-39ba2e54572f req-ee6333ca-178e-4136-8bbf-c95822c64e95 service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Neutron deleted interface 00b15ed3-61dc-432b-999a-710e050fb689; detaching it from the instance and deleting it from the info cache [ 974.953528] env[61768]: DEBUG nova.network.neutron [req-77c37dad-a790-4219-ab58-39ba2e54572f req-ee6333ca-178e-4136-8bbf-c95822c64e95 service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.046134] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.699s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.049009] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.360s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.050457] env[61768]: INFO nova.compute.claims [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 975.066337] env[61768]: INFO nova.scheduler.client.report [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted allocations for instance 271fe080-b8e9-4141-b2f3-4c4efaad81c7 [ 975.367555] env[61768]: DEBUG nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.390474] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229328, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.392909] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.393190] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.393361] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.393555] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.393710] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.393867] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.394091] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.394264] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.394440] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.394611] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.394791] env[61768]: DEBUG nova.virt.hardware [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.395588] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-743d2dea-9806-46a9-b296-a0a2ebfcba8b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.402682] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e69dda-5a66-47b1-896d-142af999f0a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.432215] env[61768]: DEBUG nova.network.neutron [-] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.456030] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-abe514a1-9702-448f-be41-1de18fb1b997 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.464953] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfc36b9-e1ee-4cfb-a78b-765c4671cdbd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.491180] env[61768]: DEBUG nova.compute.manager [req-77c37dad-a790-4219-ab58-39ba2e54572f req-ee6333ca-178e-4136-8bbf-c95822c64e95 service nova] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Detach interface failed, port_id=00b15ed3-61dc-432b-999a-710e050fb689, reason: Instance 49723fff-d1ce-476f-ac23-bd2472762d94 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 975.575048] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5b15ae5e-86aa-46d7-a7aa-a1538b5f1e2c tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "271fe080-b8e9-4141-b2f3-4c4efaad81c7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.064s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.828989] env[61768]: DEBUG nova.network.neutron [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Successfully updated port: a845dca4-da7a-491b-a5a8-c83fb17f582b {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.898869] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229328, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.935723] env[61768]: INFO nova.compute.manager [-] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Took 1.30 seconds to deallocate network for instance. [ 976.176060] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb09a09-c480-4800-b3fa-e8bd093d334f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.185189] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b303cd-975b-4027-a227-fe3d835be080 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.218909] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8cbdd4-bda6-4ec7-b2c6-ba378f3ae6b2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.226759] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d21317-cfea-4fb9-8bb7-ad4701b82074 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.240372] env[61768]: DEBUG nova.compute.provider_tree [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.332616] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.332616] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.332880] env[61768]: DEBUG nova.network.neutron [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 976.389621] env[61768]: DEBUG oslo_vmware.api [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229328, 'name': PowerOnVM_Task, 'duration_secs': 1.358428} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.390122] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 976.390379] env[61768]: INFO nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Took 9.38 seconds to spawn the instance on the hypervisor. [ 976.390579] env[61768]: DEBUG nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.391480] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a336b0-3717-4711-a3a4-4f0b6936a2e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.445965] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.605802] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.606150] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.743589] env[61768]: DEBUG nova.scheduler.client.report [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.862947] env[61768]: DEBUG nova.network.neutron [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 976.910937] env[61768]: INFO nova.compute.manager [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Took 22.23 seconds to build instance. [ 976.979379] env[61768]: DEBUG nova.compute.manager [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Received event network-vif-plugged-a845dca4-da7a-491b-a5a8-c83fb17f582b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.979608] env[61768]: DEBUG oslo_concurrency.lockutils [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] Acquiring lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.979817] env[61768]: DEBUG oslo_concurrency.lockutils [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.980044] env[61768]: DEBUG oslo_concurrency.lockutils [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.980251] env[61768]: DEBUG nova.compute.manager [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] No waiting events found dispatching network-vif-plugged-a845dca4-da7a-491b-a5a8-c83fb17f582b {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 976.980427] env[61768]: WARNING nova.compute.manager [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Received unexpected event network-vif-plugged-a845dca4-da7a-491b-a5a8-c83fb17f582b for instance with vm_state building and task_state spawning. [ 976.980596] env[61768]: DEBUG nova.compute.manager [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Received event network-changed-a845dca4-da7a-491b-a5a8-c83fb17f582b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 976.980781] env[61768]: DEBUG nova.compute.manager [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Refreshing instance network info cache due to event network-changed-a845dca4-da7a-491b-a5a8-c83fb17f582b. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 976.981015] env[61768]: DEBUG oslo_concurrency.lockutils [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] Acquiring lock "refresh_cache-9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.002800] env[61768]: DEBUG nova.network.neutron [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Updating instance_info_cache with network_info: [{"id": "a845dca4-da7a-491b-a5a8-c83fb17f582b", "address": "fa:16:3e:88:5b:10", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa845dca4-da", "ovs_interfaceid": "a845dca4-da7a-491b-a5a8-c83fb17f582b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.108832] env[61768]: DEBUG nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 977.248058] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.199s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.248644] env[61768]: DEBUG nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.251502] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.806s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.251644] env[61768]: DEBUG nova.objects.instance [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lazy-loading 'resources' on Instance uuid 49723fff-d1ce-476f-ac23-bd2472762d94 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.412600] env[61768]: DEBUG oslo_concurrency.lockutils [None req-4ad7b2e2-1cba-4330-aab3-9d42420a59ab tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.739s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.504890] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.505283] env[61768]: DEBUG nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Instance network_info: |[{"id": "a845dca4-da7a-491b-a5a8-c83fb17f582b", "address": "fa:16:3e:88:5b:10", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa845dca4-da", "ovs_interfaceid": "a845dca4-da7a-491b-a5a8-c83fb17f582b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.505579] env[61768]: DEBUG oslo_concurrency.lockutils [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] Acquired lock "refresh_cache-9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.505773] env[61768]: DEBUG nova.network.neutron [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Refreshing network info cache for port a845dca4-da7a-491b-a5a8-c83fb17f582b {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 977.506961] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:5b:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a845dca4-da7a-491b-a5a8-c83fb17f582b', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.515050] env[61768]: DEBUG oslo.service.loopingcall [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.517827] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 977.518324] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-599d83a6-04f7-441b-8862-ce4a2afe31d9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.538760] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.538760] env[61768]: value = "task-1229329" [ 977.538760] env[61768]: _type = "Task" [ 977.538760] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.546202] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229329, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.635315] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.739788] env[61768]: DEBUG nova.network.neutron [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Updated VIF entry in instance network info cache for port a845dca4-da7a-491b-a5a8-c83fb17f582b. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 977.740294] env[61768]: DEBUG nova.network.neutron [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Updating instance_info_cache with network_info: [{"id": "a845dca4-da7a-491b-a5a8-c83fb17f582b", "address": "fa:16:3e:88:5b:10", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa845dca4-da", "ovs_interfaceid": "a845dca4-da7a-491b-a5a8-c83fb17f582b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.754615] env[61768]: DEBUG nova.compute.utils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.756041] env[61768]: DEBUG nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.756203] env[61768]: DEBUG nova.network.neutron [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 977.801144] env[61768]: DEBUG nova.policy [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2280e9e6bde740c3aca30ebf70c7f73d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc444d04ff864300a45a55a3c67507e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.864857] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372615bc-d23b-46b4-8a09-645a0de41706 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.872074] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8768ffdf-299b-472e-9389-65cbb0969027 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.902749] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec40d350-d4b2-453b-848a-9b2e3627ed6c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.909641] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-635a80c0-8cb0-442a-b40f-8e74e8689f18 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.923086] env[61768]: DEBUG nova.compute.provider_tree [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.052277] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229329, 'name': CreateVM_Task, 'duration_secs': 0.291756} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.052277] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 978.052277] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.052277] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.052626] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 978.052794] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb3a0b40-970d-4b11-9a9e-3f9c6a5ee5a2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.057855] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 978.057855] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524df1b0-d3ac-5b77-96a6-720da06f59bf" [ 978.057855] env[61768]: _type = "Task" [ 978.057855] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.067429] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524df1b0-d3ac-5b77-96a6-720da06f59bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.074851] env[61768]: DEBUG nova.network.neutron [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Successfully created port: 09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.126673] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "750bc93c-65aa-4afd-a07b-cbeda0acae24" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.127049] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.127212] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "750bc93c-65aa-4afd-a07b-cbeda0acae24-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.127412] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.127594] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.129879] env[61768]: INFO nova.compute.manager [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Terminating instance [ 978.131665] env[61768]: DEBUG nova.compute.manager [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 978.131889] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 978.132737] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8777ffaa-e250-446f-9cfd-ff10095c1c47 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.140545] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 978.140839] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d18bf10-010b-4d0e-854e-cc7b12ef775a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.147407] env[61768]: DEBUG oslo_vmware.api [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 978.147407] env[61768]: value = "task-1229330" [ 978.147407] env[61768]: _type = "Task" [ 978.147407] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.155823] env[61768]: DEBUG oslo_vmware.api [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.243810] env[61768]: DEBUG oslo_concurrency.lockutils [req-d1c8e89e-c9bd-46ec-bda9-367ec8f8883b req-112fb11b-89b0-49cd-a52c-7361f9fd5633 service nova] Releasing lock "refresh_cache-9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.261842] env[61768]: DEBUG nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.425350] env[61768]: DEBUG nova.scheduler.client.report [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.568540] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]524df1b0-d3ac-5b77-96a6-720da06f59bf, 'name': SearchDatastore_Task, 'duration_secs': 0.013126} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.569269] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.569533] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 978.569757] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.569934] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.570154] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 978.570433] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab086e4e-3da2-456f-a6c1-cb8205d71a41 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.578353] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 978.578546] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 978.579304] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76f4bb34-c53f-45f2-b3b4-b375ddbcec73 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.584363] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 978.584363] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52114701-8a8a-d8f9-3d4c-b4ec8fdb3b8a" [ 978.584363] env[61768]: _type = "Task" [ 978.584363] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.594050] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52114701-8a8a-d8f9-3d4c-b4ec8fdb3b8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.658979] env[61768]: DEBUG oslo_vmware.api [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229330, 'name': PowerOffVM_Task, 'duration_secs': 0.188582} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.659301] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 978.659460] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 978.659715] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0cc38d97-8e1d-4f30-b91e-869a33a3a7f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.719113] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 978.719383] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 978.719582] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Deleting the datastore file [datastore1] 750bc93c-65aa-4afd-a07b-cbeda0acae24 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.719855] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-17727608-c7d2-4ccd-8b12-b0e1651bbd22 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.726370] env[61768]: DEBUG oslo_vmware.api [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for the task: (returnval){ [ 978.726370] env[61768]: value = "task-1229332" [ 978.726370] env[61768]: _type = "Task" [ 978.726370] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.734396] env[61768]: DEBUG oslo_vmware.api [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229332, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.931280] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.680s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.933540] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.298s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.935521] env[61768]: INFO nova.compute.claims [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 978.958554] env[61768]: INFO nova.scheduler.client.report [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Deleted allocations for instance 49723fff-d1ce-476f-ac23-bd2472762d94 [ 979.094060] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52114701-8a8a-d8f9-3d4c-b4ec8fdb3b8a, 'name': SearchDatastore_Task, 'duration_secs': 0.008635} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.094774] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c2a37d0-6a67-46ca-9043-a74827c84d16 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.099726] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 979.099726] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52894fc8-72cf-209d-7214-71c12eabccea" [ 979.099726] env[61768]: _type = "Task" [ 979.099726] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.107074] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52894fc8-72cf-209d-7214-71c12eabccea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.235681] env[61768]: DEBUG oslo_vmware.api [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Task: {'id': task-1229332, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150122} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.235950] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.236180] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 979.236371] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 979.236554] env[61768]: INFO nova.compute.manager [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Took 1.10 seconds to destroy the instance on the hypervisor. [ 979.236807] env[61768]: DEBUG oslo.service.loopingcall [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.237022] env[61768]: DEBUG nova.compute.manager [-] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.237121] env[61768]: DEBUG nova.network.neutron [-] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 979.271572] env[61768]: DEBUG nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.296995] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.297300] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.297472] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.297668] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.297826] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.297987] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.298226] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.298397] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.298573] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.298755] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.298930] env[61768]: DEBUG nova.virt.hardware [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.299816] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1c4531a-5469-4119-aea1-1deb91216675 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.307876] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f52443-71b3-44ae-b950-067bb9d17390 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.466459] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1e23c685-c842-4d8d-b399-fa6674a463f4 tempest-ServerGroupTestJSON-179049985 tempest-ServerGroupTestJSON-179049985-project-member] Lock "49723fff-d1ce-476f-ac23-bd2472762d94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.967s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.593491] env[61768]: DEBUG nova.compute.manager [req-2049fe00-6997-4ef1-8e8f-2cadbb138c9f req-e4fe925a-1b2e-4e22-8be4-c4424a85532e service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Received event network-vif-deleted-2be36d33-ba56-45fc-b047-be0c646b844d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.597122] env[61768]: INFO nova.compute.manager [req-2049fe00-6997-4ef1-8e8f-2cadbb138c9f req-e4fe925a-1b2e-4e22-8be4-c4424a85532e service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Neutron deleted interface 2be36d33-ba56-45fc-b047-be0c646b844d; detaching it from the instance and deleting it from the info cache [ 979.597122] env[61768]: DEBUG nova.network.neutron [req-2049fe00-6997-4ef1-8e8f-2cadbb138c9f req-e4fe925a-1b2e-4e22-8be4-c4424a85532e service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.609680] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52894fc8-72cf-209d-7214-71c12eabccea, 'name': SearchDatastore_Task, 'duration_secs': 0.008842} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.610489] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.611177] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f/9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 979.611177] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a9c349fa-7458-47dd-bc81-9d82291dd842 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.617825] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 979.617825] env[61768]: value = "task-1229333" [ 979.617825] env[61768]: _type = "Task" [ 979.617825] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.627212] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229333, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.788035] env[61768]: DEBUG nova.network.neutron [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Successfully updated port: 09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.060082] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-965628f7-df8f-41a5-99c4-72f335930a91 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.069677] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dadb4776-2c5d-482d-9475-a30e066c62f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.073124] env[61768]: DEBUG nova.network.neutron [-] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.102864] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e7d83a3c-7935-4f51-b734-d911514007fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.105449] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f54712e6-b4a4-44ea-bf75-2bc83d8e6cd2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.113970] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6068f2b8-5b95-4dd9-8677-b67b9c30ce31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.120857] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2df25e5a-84e6-4e7b-962f-e06b6511ac71 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.144614] env[61768]: DEBUG nova.compute.provider_tree [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 980.153030] env[61768]: DEBUG nova.compute.manager [req-2049fe00-6997-4ef1-8e8f-2cadbb138c9f req-e4fe925a-1b2e-4e22-8be4-c4424a85532e service nova] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Detach interface failed, port_id=2be36d33-ba56-45fc-b047-be0c646b844d, reason: Instance 750bc93c-65aa-4afd-a07b-cbeda0acae24 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.154109] env[61768]: DEBUG nova.scheduler.client.report [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.160430] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229333, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471416} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.161476] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f/9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 980.161476] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 980.161476] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae3c0ce9-e268-42ba-8665-39596d2aa295 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.168594] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 980.168594] env[61768]: value = "task-1229334" [ 980.168594] env[61768]: _type = "Task" [ 980.168594] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.177386] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229334, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.297097] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.297097] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.297097] env[61768]: DEBUG nova.network.neutron [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.576324] env[61768]: INFO nova.compute.manager [-] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Took 1.34 seconds to deallocate network for instance. [ 980.661847] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.728s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.662549] env[61768]: DEBUG nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 980.677394] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229334, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062172} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.677676] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.678469] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83da32bc-d81f-4553-a29d-3d8d97c13e72 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.702628] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f/9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.703728] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0f3571b9-e0d7-4ff6-9a60-9c6ee1701fff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.725107] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 980.725107] env[61768]: value = "task-1229335" [ 980.725107] env[61768]: _type = "Task" [ 980.725107] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.733296] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229335, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.858893] env[61768]: DEBUG nova.network.neutron [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 981.083292] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.083601] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.083840] env[61768]: DEBUG nova.objects.instance [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lazy-loading 'resources' on Instance uuid 750bc93c-65aa-4afd-a07b-cbeda0acae24 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.103293] env[61768]: DEBUG nova.network.neutron [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updating instance_info_cache with network_info: [{"id": "09109871-16a8-42aa-be6d-e34e9885e21d", "address": "fa:16:3e:85:c9:fd", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09109871-16", "ovs_interfaceid": "09109871-16a8-42aa-be6d-e34e9885e21d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.167429] env[61768]: DEBUG nova.compute.utils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 981.168881] env[61768]: DEBUG nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 981.169078] env[61768]: DEBUG nova.network.neutron [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 981.209162] env[61768]: DEBUG nova.policy [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '65fd5852bc16429aa2904b4351c8249d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c7857b5217cc499ca8c9c3b8a5471db7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 981.235270] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229335, 'name': ReconfigVM_Task, 'duration_secs': 0.306226} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.235545] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f/9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.236202] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5f22f44d-da71-4340-aac1-12b150572a60 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.242585] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 981.242585] env[61768]: value = "task-1229336" [ 981.242585] env[61768]: _type = "Task" [ 981.242585] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.252495] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229336, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.538797] env[61768]: DEBUG nova.network.neutron [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Successfully created port: 127c5490-87ca-460f-9891-b309eae19b78 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 981.605338] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.605685] env[61768]: DEBUG nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Instance network_info: |[{"id": "09109871-16a8-42aa-be6d-e34e9885e21d", "address": "fa:16:3e:85:c9:fd", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09109871-16", "ovs_interfaceid": "09109871-16a8-42aa-be6d-e34e9885e21d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 981.608555] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:c9:fd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13e71dbb-4279-427c-b39d-ba5df9895e58', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '09109871-16a8-42aa-be6d-e34e9885e21d', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.616895] env[61768]: DEBUG oslo.service.loopingcall [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.617378] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 981.617613] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6eefdb5e-959c-48fd-af98-c5cfecb26594 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.639251] env[61768]: DEBUG nova.compute.manager [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Received event network-vif-plugged-09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.639510] env[61768]: DEBUG oslo_concurrency.lockutils [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] Acquiring lock "86f41ef0-4c92-4626-bb83-1148144c597c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.639767] env[61768]: DEBUG oslo_concurrency.lockutils [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] Lock "86f41ef0-4c92-4626-bb83-1148144c597c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.639989] env[61768]: DEBUG oslo_concurrency.lockutils [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] Lock "86f41ef0-4c92-4626-bb83-1148144c597c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.640320] env[61768]: DEBUG nova.compute.manager [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] No waiting events found dispatching network-vif-plugged-09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.640441] env[61768]: WARNING nova.compute.manager [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Received unexpected event network-vif-plugged-09109871-16a8-42aa-be6d-e34e9885e21d for instance with vm_state building and task_state spawning. [ 981.640652] env[61768]: DEBUG nova.compute.manager [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Received event network-changed-09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.640883] env[61768]: DEBUG nova.compute.manager [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Refreshing instance network info cache due to event network-changed-09109871-16a8-42aa-be6d-e34e9885e21d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.641228] env[61768]: DEBUG oslo_concurrency.lockutils [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] Acquiring lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.641319] env[61768]: DEBUG oslo_concurrency.lockutils [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] Acquired lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.641515] env[61768]: DEBUG nova.network.neutron [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Refreshing network info cache for port 09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.649183] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.649183] env[61768]: value = "task-1229337" [ 981.649183] env[61768]: _type = "Task" [ 981.649183] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.662276] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229337, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.672725] env[61768]: DEBUG nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 981.714592] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4861ff51-f176-412b-a27f-89a59dddc82e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.723693] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45548e9e-6a59-4dcb-98e3-2a7d85f11cf8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.761447] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e8f2e2-a3bb-4f88-b98c-8e94a7fd3bc4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.769551] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229336, 'name': Rename_Task, 'duration_secs': 0.138723} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.771701] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 981.771993] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e1877752-798b-44c6-a88b-13bc3956e9be {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.774415] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f5e9589-24ea-4f86-8dcf-8258ae283e14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.791078] env[61768]: DEBUG nova.compute.provider_tree [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.794016] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 981.794016] env[61768]: value = "task-1229338" [ 981.794016] env[61768]: _type = "Task" [ 981.794016] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.802648] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229338, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.161768] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229337, 'name': CreateVM_Task, 'duration_secs': 0.324153} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.161768] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 982.162151] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.162151] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.162417] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.162686] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-95568193-4ce8-4fb0-b122-e680362b56bb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.168140] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 982.168140] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529761ee-a360-fcb2-0b7d-26e24611fef9" [ 982.168140] env[61768]: _type = "Task" [ 982.168140] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.180910] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529761ee-a360-fcb2-0b7d-26e24611fef9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.294218] env[61768]: DEBUG nova.scheduler.client.report [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.308744] env[61768]: DEBUG oslo_vmware.api [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229338, 'name': PowerOnVM_Task, 'duration_secs': 0.435018} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.309108] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 982.309349] env[61768]: INFO nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Took 6.94 seconds to spawn the instance on the hypervisor. [ 982.309562] env[61768]: DEBUG nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.310362] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3ffd56-0a60-4996-80f7-42fb0c19df39 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.369036] env[61768]: DEBUG nova.network.neutron [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updated VIF entry in instance network info cache for port 09109871-16a8-42aa-be6d-e34e9885e21d. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.369469] env[61768]: DEBUG nova.network.neutron [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updating instance_info_cache with network_info: [{"id": "09109871-16a8-42aa-be6d-e34e9885e21d", "address": "fa:16:3e:85:c9:fd", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09109871-16", "ovs_interfaceid": "09109871-16a8-42aa-be6d-e34e9885e21d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.678254] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529761ee-a360-fcb2-0b7d-26e24611fef9, 'name': SearchDatastore_Task, 'duration_secs': 0.009738} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.678567] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.679159] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.679159] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.679280] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.679400] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.679659] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afa62d15-ba6c-41d0-9864-7e233f8cd029 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.682602] env[61768]: DEBUG nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 982.688247] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.688247] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.688706] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19394aae-1541-40d4-8c4a-c7f236a76a9b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.694969] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 982.694969] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e5b460-20bd-42f9-c6de-e27933aa3a4c" [ 982.694969] env[61768]: _type = "Task" [ 982.694969] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.702218] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e5b460-20bd-42f9-c6de-e27933aa3a4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.710538] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 982.710949] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 982.711177] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 982.711386] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 982.711543] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 982.711699] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 982.711972] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 982.712185] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 982.712368] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 982.712545] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 982.712726] env[61768]: DEBUG nova.virt.hardware [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 982.713516] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13c1bff1-7c24-4f2e-b8d2-c205c190d5d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.720696] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2977d3d7-4c0f-41cb-b11b-7420fc28a435 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.804793] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.827691] env[61768]: INFO nova.scheduler.client.report [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Deleted allocations for instance 750bc93c-65aa-4afd-a07b-cbeda0acae24 [ 982.829273] env[61768]: INFO nova.compute.manager [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Took 17.16 seconds to build instance. [ 982.872419] env[61768]: DEBUG oslo_concurrency.lockutils [req-b018b153-8f0d-4ca8-9932-0d5f0eed2fbc req-b9c0f594-7982-4203-abbd-6ce4cca72752 service nova] Releasing lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.209414] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52e5b460-20bd-42f9-c6de-e27933aa3a4c, 'name': SearchDatastore_Task, 'duration_secs': 0.008845} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.210605] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0009ef65-e912-4406-9d74-84c252bb3bbd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.217223] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 983.217223] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a7b2ff-a6ad-0f52-681f-1ffb4f3cce57" [ 983.217223] env[61768]: _type = "Task" [ 983.217223] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.227530] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a7b2ff-a6ad-0f52-681f-1ffb4f3cce57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.331547] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7aa36179-3ead-4cc9-a478-45615e96c49e tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.686s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.337298] env[61768]: DEBUG oslo_concurrency.lockutils [None req-30b67db0-3e25-4a34-b097-28e0b1325c46 tempest-InstanceActionsNegativeTestJSON-1757049735 tempest-InstanceActionsNegativeTestJSON-1757049735-project-member] Lock "750bc93c-65aa-4afd-a07b-cbeda0acae24" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.210s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.390808] env[61768]: DEBUG nova.network.neutron [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Successfully updated port: 127c5490-87ca-460f-9891-b309eae19b78 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 983.532711] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efeec17-5ce3-47eb-8a71-f82d4ad13e9b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.538769] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bccee0eb-bbae-4e17-8dd6-f6e416067d86 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Suspending the VM {{(pid=61768) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 983.539065] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e881c1f9-cc95-48bb-aabb-0e3a8eed1ad5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.545344] env[61768]: DEBUG oslo_vmware.api [None req-bccee0eb-bbae-4e17-8dd6-f6e416067d86 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 983.545344] env[61768]: value = "task-1229339" [ 983.545344] env[61768]: _type = "Task" [ 983.545344] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.555889] env[61768]: DEBUG oslo_vmware.api [None req-bccee0eb-bbae-4e17-8dd6-f6e416067d86 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229339, 'name': SuspendVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.672050] env[61768]: DEBUG nova.compute.manager [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Received event network-vif-plugged-127c5490-87ca-460f-9891-b309eae19b78 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.672328] env[61768]: DEBUG oslo_concurrency.lockutils [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] Acquiring lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.672614] env[61768]: DEBUG oslo_concurrency.lockutils [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.672818] env[61768]: DEBUG oslo_concurrency.lockutils [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.672965] env[61768]: DEBUG nova.compute.manager [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] No waiting events found dispatching network-vif-plugged-127c5490-87ca-460f-9891-b309eae19b78 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 983.673334] env[61768]: WARNING nova.compute.manager [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Received unexpected event network-vif-plugged-127c5490-87ca-460f-9891-b309eae19b78 for instance with vm_state building and task_state spawning. [ 983.673526] env[61768]: DEBUG nova.compute.manager [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Received event network-changed-127c5490-87ca-460f-9891-b309eae19b78 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.673716] env[61768]: DEBUG nova.compute.manager [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Refreshing instance network info cache due to event network-changed-127c5490-87ca-460f-9891-b309eae19b78. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.673890] env[61768]: DEBUG oslo_concurrency.lockutils [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] Acquiring lock "refresh_cache-e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.674048] env[61768]: DEBUG oslo_concurrency.lockutils [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] Acquired lock "refresh_cache-e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.674219] env[61768]: DEBUG nova.network.neutron [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Refreshing network info cache for port 127c5490-87ca-460f-9891-b309eae19b78 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 983.728598] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a7b2ff-a6ad-0f52-681f-1ffb4f3cce57, 'name': SearchDatastore_Task, 'duration_secs': 0.009349} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.728986] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.729291] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 86f41ef0-4c92-4626-bb83-1148144c597c/86f41ef0-4c92-4626-bb83-1148144c597c.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.729568] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-66fcc4ab-8da3-4252-8e99-3f9646d3cd41 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.736623] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 983.736623] env[61768]: value = "task-1229340" [ 983.736623] env[61768]: _type = "Task" [ 983.736623] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.744579] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229340, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.896889] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "refresh_cache-e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 984.060889] env[61768]: DEBUG oslo_vmware.api [None req-bccee0eb-bbae-4e17-8dd6-f6e416067d86 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229339, 'name': SuspendVM_Task} progress is 62%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.228116] env[61768]: DEBUG nova.network.neutron [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 984.247974] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229340, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463157} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.248315] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 86f41ef0-4c92-4626-bb83-1148144c597c/86f41ef0-4c92-4626-bb83-1148144c597c.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 984.248550] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.248872] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-608b6dca-4c8d-4bd7-9784-5dd7b62100cd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.257069] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 984.257069] env[61768]: value = "task-1229341" [ 984.257069] env[61768]: _type = "Task" [ 984.257069] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.268726] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229341, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.344270] env[61768]: DEBUG nova.network.neutron [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 984.561249] env[61768]: DEBUG oslo_vmware.api [None req-bccee0eb-bbae-4e17-8dd6-f6e416067d86 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229339, 'name': SuspendVM_Task, 'duration_secs': 0.645318} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.561341] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-bccee0eb-bbae-4e17-8dd6-f6e416067d86 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Suspended the VM {{(pid=61768) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 984.561513] env[61768]: DEBUG nova.compute.manager [None req-bccee0eb-bbae-4e17-8dd6-f6e416067d86 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.562341] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e89bfc-2242-4799-b134-667168c893b3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.769255] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229341, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064517} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.769877] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.771088] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e9a74b-b9f2-4aa6-b6cf-fcb41c6c5185 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.799517] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 86f41ef0-4c92-4626-bb83-1148144c597c/86f41ef0-4c92-4626-bb83-1148144c597c.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.800462] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbf9183c-a582-4878-b84a-78a98da455ae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.828538] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 984.828538] env[61768]: value = "task-1229342" [ 984.828538] env[61768]: _type = "Task" [ 984.828538] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.835235] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229342, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.850362] env[61768]: DEBUG oslo_concurrency.lockutils [req-72661ef8-133a-451a-b69b-f176547f192c req-ad096afe-db02-4f07-a25b-d10b88dc1a97 service nova] Releasing lock "refresh_cache-e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.850362] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "refresh_cache-e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.850362] env[61768]: DEBUG nova.network.neutron [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 985.039599] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.040017] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.335912] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229342, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.402940] env[61768]: DEBUG nova.network.neutron [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 985.543068] env[61768]: DEBUG nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 985.582934] env[61768]: DEBUG nova.network.neutron [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Updating instance_info_cache with network_info: [{"id": "127c5490-87ca-460f-9891-b309eae19b78", "address": "fa:16:3e:c3:84:e6", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap127c5490-87", "ovs_interfaceid": "127c5490-87ca-460f-9891-b309eae19b78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.684088] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "aa568a72-afba-4a25-a048-ca62191f9212" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 985.684088] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "aa568a72-afba-4a25-a048-ca62191f9212" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.836025] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229342, 'name': ReconfigVM_Task, 'duration_secs': 0.662175} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.836662] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 86f41ef0-4c92-4626-bb83-1148144c597c/86f41ef0-4c92-4626-bb83-1148144c597c.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.836936] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c2fcd25d-acc0-434c-8fd4-655f94d71220 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.843327] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 985.843327] env[61768]: value = "task-1229343" [ 985.843327] env[61768]: _type = "Task" [ 985.843327] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.852577] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229343, 'name': Rename_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.063277] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.063619] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.065454] env[61768]: INFO nova.compute.claims [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 986.086127] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "refresh_cache-e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.086448] env[61768]: DEBUG nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Instance network_info: |[{"id": "127c5490-87ca-460f-9891-b309eae19b78", "address": "fa:16:3e:c3:84:e6", "network": {"id": "759833b8-c22b-42d4-af6f-38766dbf9fa6", "bridge": "br-int", "label": "tempest-ServersTestJSON-317747395-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c7857b5217cc499ca8c9c3b8a5471db7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2f204ad5-8a45-4372-80ba-010fe0f9a337", "external-id": "nsx-vlan-transportzone-593", "segmentation_id": 593, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap127c5490-87", "ovs_interfaceid": "127c5490-87ca-460f-9891-b309eae19b78", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 986.086893] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:84:e6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2f204ad5-8a45-4372-80ba-010fe0f9a337', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '127c5490-87ca-460f-9891-b309eae19b78', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 986.095829] env[61768]: DEBUG oslo.service.loopingcall [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.096092] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 986.096378] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b3e39225-41c3-4305-9838-383fb84dfc50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.112747] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.112974] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.113196] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.113388] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.113560] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.118170] env[61768]: INFO nova.compute.manager [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Terminating instance [ 986.119958] env[61768]: DEBUG nova.compute.manager [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.120094] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 986.120873] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56f6a3b3-8117-4ab8-a7dc-747dc143473a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.125485] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 986.125485] env[61768]: value = "task-1229344" [ 986.125485] env[61768]: _type = "Task" [ 986.125485] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.130660] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 986.131425] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-53f6a814-66a0-4b61-a1b3-d972f327bf70 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.135993] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229344, 'name': CreateVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.187668] env[61768]: DEBUG nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 986.195753] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 986.196086] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 986.196427] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleting the datastore file [datastore2] 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.196854] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad1682f3-8b1e-4071-89fc-f885e3776a4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.205385] env[61768]: DEBUG oslo_vmware.api [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 986.205385] env[61768]: value = "task-1229346" [ 986.205385] env[61768]: _type = "Task" [ 986.205385] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.215886] env[61768]: DEBUG oslo_vmware.api [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.354028] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229343, 'name': Rename_Task, 'duration_secs': 0.138962} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.354440] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.354734] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ed826ab4-3d28-4430-be57-93186453edae {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.361407] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 986.361407] env[61768]: value = "task-1229347" [ 986.361407] env[61768]: _type = "Task" [ 986.361407] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.369408] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229347, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.497329] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.497591] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.636615] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229344, 'name': CreateVM_Task, 'duration_secs': 0.291228} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.636839] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 986.637603] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.637874] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.638316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 986.638625] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3b399e9-45a9-46b0-9141-2bee56549a5a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.643474] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 986.643474] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5219c1bd-5b7d-5f6f-ac5b-1ba5888b4155" [ 986.643474] env[61768]: _type = "Task" [ 986.643474] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.653014] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5219c1bd-5b7d-5f6f-ac5b-1ba5888b4155, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.711825] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.715598] env[61768]: DEBUG oslo_vmware.api [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.309338} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.715886] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 986.716129] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 986.716356] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 986.716572] env[61768]: INFO nova.compute.manager [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Took 0.60 seconds to destroy the instance on the hypervisor. [ 986.716852] env[61768]: DEBUG oslo.service.loopingcall [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 986.717086] env[61768]: DEBUG nova.compute.manager [-] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 986.717206] env[61768]: DEBUG nova.network.neutron [-] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 986.872756] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229347, 'name': PowerOnVM_Task} progress is 91%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.003679] env[61768]: DEBUG nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 987.044738] env[61768]: DEBUG nova.compute.manager [req-efa903dd-1496-4f77-a649-5ac67fdaffe2 req-5433b2f7-fc3d-45d1-b7df-39b68e42f1f4 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Received event network-vif-deleted-a845dca4-da7a-491b-a5a8-c83fb17f582b {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.045237] env[61768]: INFO nova.compute.manager [req-efa903dd-1496-4f77-a649-5ac67fdaffe2 req-5433b2f7-fc3d-45d1-b7df-39b68e42f1f4 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Neutron deleted interface a845dca4-da7a-491b-a5a8-c83fb17f582b; detaching it from the instance and deleting it from the info cache [ 987.045237] env[61768]: DEBUG nova.network.neutron [req-efa903dd-1496-4f77-a649-5ac67fdaffe2 req-5433b2f7-fc3d-45d1-b7df-39b68e42f1f4 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.155728] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5219c1bd-5b7d-5f6f-ac5b-1ba5888b4155, 'name': SearchDatastore_Task, 'duration_secs': 0.009855} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.155901] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.156154] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 987.156397] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.156560] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.156746] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 987.157015] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b2e8287-c1ac-461b-860a-52405140421f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.173629] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 987.173761] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 987.176612] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ede1b960-b628-4dc4-93b7-b3b00f97116c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.182795] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 987.182795] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52892f40-3f14-7342-87db-9f5eab8372f1" [ 987.182795] env[61768]: _type = "Task" [ 987.182795] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.192735] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52892f40-3f14-7342-87db-9f5eab8372f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.200063] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0263f0bd-3fc1-4a9d-90e9-e8c1050005ba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.206739] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c211f750-2adc-4dfb-9d6f-b3874888d21e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.238356] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8398428-ea46-4f83-84d4-9e8b488048ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.245754] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89502a0-9b25-4f6f-8988-f2ccc00c0a0c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.258737] env[61768]: DEBUG nova.compute.provider_tree [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.372144] env[61768]: DEBUG oslo_vmware.api [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229347, 'name': PowerOnVM_Task, 'duration_secs': 0.550358} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.372468] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.372623] env[61768]: INFO nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Took 8.10 seconds to spawn the instance on the hypervisor. [ 987.372810] env[61768]: DEBUG nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.373568] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a4a689-14d1-493f-93b7-3c3e5574078a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.522263] env[61768]: DEBUG nova.network.neutron [-] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.526515] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.548025] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c0c79bb-d660-4822-a87b-afaec6b6f197 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.557054] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6dc33d-efba-4fd3-a723-cf076a8bc3a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.583577] env[61768]: DEBUG nova.compute.manager [req-efa903dd-1496-4f77-a649-5ac67fdaffe2 req-5433b2f7-fc3d-45d1-b7df-39b68e42f1f4 service nova] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Detach interface failed, port_id=a845dca4-da7a-491b-a5a8-c83fb17f582b, reason: Instance 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 987.693701] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52892f40-3f14-7342-87db-9f5eab8372f1, 'name': SearchDatastore_Task, 'duration_secs': 0.023281} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.694550] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e31b2b2-68e4-47f7-ac4e-8f31e827279e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.699675] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 987.699675] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528764da-dd8c-94e0-0811-fa6cec6fea0c" [ 987.699675] env[61768]: _type = "Task" [ 987.699675] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.707029] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528764da-dd8c-94e0-0811-fa6cec6fea0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.761405] env[61768]: DEBUG nova.scheduler.client.report [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.890724] env[61768]: INFO nova.compute.manager [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Took 15.21 seconds to build instance. [ 988.025277] env[61768]: INFO nova.compute.manager [-] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Took 1.31 seconds to deallocate network for instance. [ 988.210995] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528764da-dd8c-94e0-0811-fa6cec6fea0c, 'name': SearchDatastore_Task, 'duration_secs': 0.012705} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.211484] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.211772] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa/e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 988.212153] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-92397f70-a52c-46e8-83e5-b4157a414699 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.219062] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 988.219062] env[61768]: value = "task-1229348" [ 988.219062] env[61768]: _type = "Task" [ 988.219062] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.227642] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229348, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.268207] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.203s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.268207] env[61768]: DEBUG nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 988.269908] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.558s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.271478] env[61768]: INFO nova.compute.claims [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 988.392690] env[61768]: DEBUG oslo_concurrency.lockutils [None req-442dcb22-dae3-4f7b-8a60-d662a50f8066 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.724s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.532132] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.731217] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229348, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.504135} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.731609] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa/e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 988.731923] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 988.732278] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-964e6471-fe49-4242-8bef-f0b8a181f7cd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.738313] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 988.738313] env[61768]: value = "task-1229349" [ 988.738313] env[61768]: _type = "Task" [ 988.738313] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.746445] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229349, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.776130] env[61768]: DEBUG nova.compute.utils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 988.777796] env[61768]: DEBUG nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 988.777796] env[61768]: DEBUG nova.network.neutron [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 988.821494] env[61768]: DEBUG nova.policy [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26e3d09d20274851950ae324bcf85a14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4c2ff63d9d94a41b0dbc7d597e762a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 989.076638] env[61768]: DEBUG nova.compute.manager [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Received event network-changed-09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.076880] env[61768]: DEBUG nova.compute.manager [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Refreshing instance network info cache due to event network-changed-09109871-16a8-42aa-be6d-e34e9885e21d. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.077164] env[61768]: DEBUG oslo_concurrency.lockutils [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] Acquiring lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.077341] env[61768]: DEBUG oslo_concurrency.lockutils [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] Acquired lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.077527] env[61768]: DEBUG nova.network.neutron [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Refreshing network info cache for port 09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.102440] env[61768]: DEBUG nova.network.neutron [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Successfully created port: b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 989.248559] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229349, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067862} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.248559] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 989.249090] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc7d2c7-d381-4323-bc80-73f73b59b8af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.270677] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa/e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 989.270992] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23d9cf6b-2db4-4736-9c87-724de071c662 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.285474] env[61768]: DEBUG nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 989.296928] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 989.296928] env[61768]: value = "task-1229350" [ 989.296928] env[61768]: _type = "Task" [ 989.296928] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.306880] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229350, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.427875] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2ead5a-edf1-45e1-8373-bb91e05198bf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.438055] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f40e6f-e7e5-4620-b907-68e124177e13 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.468601] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75627190-ccdb-4c7c-982f-09481ff8bfe2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.476634] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a50024-8e22-4f1d-ac33-82766bf4d2ea {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.491510] env[61768]: DEBUG nova.compute.provider_tree [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.789132] env[61768]: DEBUG nova.network.neutron [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updated VIF entry in instance network info cache for port 09109871-16a8-42aa-be6d-e34e9885e21d. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 989.789537] env[61768]: DEBUG nova.network.neutron [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updating instance_info_cache with network_info: [{"id": "09109871-16a8-42aa-be6d-e34e9885e21d", "address": "fa:16:3e:85:c9:fd", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.144", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap09109871-16", "ovs_interfaceid": "09109871-16a8-42aa-be6d-e34e9885e21d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.807175] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229350, 'name': ReconfigVM_Task, 'duration_secs': 0.28958} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.807468] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Reconfigured VM instance instance-00000068 to attach disk [datastore1] e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa/e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 989.808131] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6ceb2019-88f4-4a04-97ba-b3d042754fff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.815251] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 989.815251] env[61768]: value = "task-1229351" [ 989.815251] env[61768]: _type = "Task" [ 989.815251] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.823170] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229351, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.994445] env[61768]: DEBUG nova.scheduler.client.report [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 990.292461] env[61768]: DEBUG oslo_concurrency.lockutils [req-f8f9d1c6-ea73-43f5-b6db-fc8472981646 req-8fc94064-63f2-4838-b3fd-93ccff7beb3b service nova] Releasing lock "refresh_cache-86f41ef0-4c92-4626-bb83-1148144c597c" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.295795] env[61768]: DEBUG nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 990.321640] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 990.321893] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 990.322074] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 990.322274] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 990.322429] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 990.322584] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 990.322795] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 990.322965] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 990.323155] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 990.323324] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 990.323503] env[61768]: DEBUG nova.virt.hardware [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 990.324284] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db8615e-6e7f-48b1-bbdf-58dc84d27995 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.332183] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229351, 'name': Rename_Task, 'duration_secs': 0.146689} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.334137] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 990.334405] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-58c2af5b-307a-4210-b9f7-f1cb21909f71 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.336644] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0e88267-1587-4c70-a3bc-02de50a0fa2b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.351083] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 990.351083] env[61768]: value = "task-1229352" [ 990.351083] env[61768]: _type = "Task" [ 990.351083] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.358868] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229352, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.502084] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.232s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.502698] env[61768]: DEBUG nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 990.505490] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.979s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.506922] env[61768]: INFO nova.compute.claims [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.632703] env[61768]: DEBUG nova.network.neutron [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Successfully updated port: b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.860789] env[61768]: DEBUG oslo_vmware.api [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229352, 'name': PowerOnVM_Task, 'duration_secs': 0.48799} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.861139] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 990.861371] env[61768]: INFO nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Took 8.18 seconds to spawn the instance on the hypervisor. [ 990.861562] env[61768]: DEBUG nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.862348] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7eb13888-c996-4984-86ee-bbf7088efb1e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.012235] env[61768]: DEBUG nova.compute.utils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 991.015443] env[61768]: DEBUG nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 991.015617] env[61768]: DEBUG nova.network.neutron [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 991.054465] env[61768]: DEBUG nova.policy [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26e3d09d20274851950ae324bcf85a14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4c2ff63d9d94a41b0dbc7d597e762a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 991.107138] env[61768]: DEBUG nova.compute.manager [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Received event network-vif-plugged-b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.107386] env[61768]: DEBUG oslo_concurrency.lockutils [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] Acquiring lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.107607] env[61768]: DEBUG oslo_concurrency.lockutils [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.107783] env[61768]: DEBUG oslo_concurrency.lockutils [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.107957] env[61768]: DEBUG nova.compute.manager [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] No waiting events found dispatching network-vif-plugged-b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.108161] env[61768]: WARNING nova.compute.manager [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Received unexpected event network-vif-plugged-b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 for instance with vm_state building and task_state spawning. [ 991.108401] env[61768]: DEBUG nova.compute.manager [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Received event network-changed-b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.108491] env[61768]: DEBUG nova.compute.manager [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Refreshing instance network info cache due to event network-changed-b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.108674] env[61768]: DEBUG oslo_concurrency.lockutils [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] Acquiring lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.108815] env[61768]: DEBUG oslo_concurrency.lockutils [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] Acquired lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.109024] env[61768]: DEBUG nova.network.neutron [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Refreshing network info cache for port b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 991.137697] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.355299] env[61768]: DEBUG nova.network.neutron [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Successfully created port: 5913b4b2-1038-4537-9cd9-b466dc48d869 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 991.379814] env[61768]: INFO nova.compute.manager [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Took 13.77 seconds to build instance. [ 991.516899] env[61768]: DEBUG nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 991.649350] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e1da3a-44d4-4f9e-bc86-0b2c1036d1ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.652902] env[61768]: DEBUG nova.network.neutron [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 991.660069] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eee218a-d44b-441a-b7cf-0eee0eb3a5ba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.690076] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e38f5067-419a-493b-bee6-b4a339250dcf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.698140] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32ee52f-c759-46f1-9426-a99cdcbed269 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.712510] env[61768]: DEBUG nova.compute.provider_tree [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.749280] env[61768]: DEBUG nova.network.neutron [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.882417] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e45641e6-8205-4e2f-910d-8f0616883680 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.276s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.217545] env[61768]: DEBUG nova.scheduler.client.report [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.251685] env[61768]: DEBUG oslo_concurrency.lockutils [req-0d48de33-77ca-47a8-9484-eb44429487ac req-f2990784-633a-4129-b1b5-98f52cfce800 service nova] Releasing lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 992.252100] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 992.252278] env[61768]: DEBUG nova.network.neutron [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 992.333332] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.333585] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.517817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.518175] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.518338] env[61768]: DEBUG nova.compute.manager [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.519243] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594cf5f7-14d6-471d-af1e-38669e20d0c7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.526333] env[61768]: DEBUG nova.compute.manager [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61768) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 992.526813] env[61768]: DEBUG nova.objects.instance [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'flavor' on Instance uuid e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.530465] env[61768]: DEBUG nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 992.554585] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.554845] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.555020] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.555213] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.555370] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.555525] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.555736] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.555905] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.556102] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.556284] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.556464] env[61768]: DEBUG nova.virt.hardware [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.557529] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58151c6a-21ba-4829-aaca-6b022f7c9b04 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.565362] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7a5cc7d-d3d3-4184-820e-446436c0e81d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.722989] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.217s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.723583] env[61768]: DEBUG nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 992.726119] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.194s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.726353] env[61768]: DEBUG nova.objects.instance [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'resources' on Instance uuid 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.800406] env[61768]: DEBUG nova.compute.manager [req-9ec14bbf-ea7a-428e-978e-30d40186f331 req-f34a047a-512a-4ab3-8c8e-fe2af92c2790 service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Received event network-vif-plugged-5913b4b2-1038-4537-9cd9-b466dc48d869 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.800644] env[61768]: DEBUG oslo_concurrency.lockutils [req-9ec14bbf-ea7a-428e-978e-30d40186f331 req-f34a047a-512a-4ab3-8c8e-fe2af92c2790 service nova] Acquiring lock "aa568a72-afba-4a25-a048-ca62191f9212-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.800861] env[61768]: DEBUG oslo_concurrency.lockutils [req-9ec14bbf-ea7a-428e-978e-30d40186f331 req-f34a047a-512a-4ab3-8c8e-fe2af92c2790 service nova] Lock "aa568a72-afba-4a25-a048-ca62191f9212-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.801115] env[61768]: DEBUG oslo_concurrency.lockutils [req-9ec14bbf-ea7a-428e-978e-30d40186f331 req-f34a047a-512a-4ab3-8c8e-fe2af92c2790 service nova] Lock "aa568a72-afba-4a25-a048-ca62191f9212-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.801314] env[61768]: DEBUG nova.compute.manager [req-9ec14bbf-ea7a-428e-978e-30d40186f331 req-f34a047a-512a-4ab3-8c8e-fe2af92c2790 service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] No waiting events found dispatching network-vif-plugged-5913b4b2-1038-4537-9cd9-b466dc48d869 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 992.801489] env[61768]: WARNING nova.compute.manager [req-9ec14bbf-ea7a-428e-978e-30d40186f331 req-f34a047a-512a-4ab3-8c8e-fe2af92c2790 service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Received unexpected event network-vif-plugged-5913b4b2-1038-4537-9cd9-b466dc48d869 for instance with vm_state building and task_state spawning. [ 992.805534] env[61768]: DEBUG nova.network.neutron [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 992.837169] env[61768]: DEBUG nova.compute.utils [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.943146] env[61768]: DEBUG nova.network.neutron [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updating instance_info_cache with network_info: [{"id": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "address": "fa:16:3e:b3:4f:18", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4f2467a-e3", "ovs_interfaceid": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.032288] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.032554] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d28fb9cb-7781-4ff3-b4d3-3d5f17a74f1d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.040928] env[61768]: DEBUG oslo_vmware.api [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 993.040928] env[61768]: value = "task-1229353" [ 993.040928] env[61768]: _type = "Task" [ 993.040928] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.048600] env[61768]: DEBUG oslo_vmware.api [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229353, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.229319] env[61768]: DEBUG nova.compute.utils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 993.234864] env[61768]: DEBUG nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Not allocating networking since 'none' was specified. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 993.341544] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.350444] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101b50f1-d044-4bd2-8c13-95a4de0e657a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.358011] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b4edcee-1485-41a4-a807-d84fc6e7bbb7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.388109] env[61768]: DEBUG nova.network.neutron [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Successfully updated port: 5913b4b2-1038-4537-9cd9-b466dc48d869 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 993.390144] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb03ead6-9aa7-4ed8-96a8-173080ded12b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.397696] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e033991b-a916-4c7a-a687-0867c21b2804 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.412321] env[61768]: DEBUG nova.compute.provider_tree [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.414617] env[61768]: DEBUG nova.compute.manager [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Received event network-changed-5913b4b2-1038-4537-9cd9-b466dc48d869 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 993.414818] env[61768]: DEBUG nova.compute.manager [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Refreshing instance network info cache due to event network-changed-5913b4b2-1038-4537-9cd9-b466dc48d869. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 993.415048] env[61768]: DEBUG oslo_concurrency.lockutils [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] Acquiring lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.415206] env[61768]: DEBUG oslo_concurrency.lockutils [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] Acquired lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 993.415373] env[61768]: DEBUG nova.network.neutron [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Refreshing network info cache for port 5913b4b2-1038-4537-9cd9-b466dc48d869 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 993.445897] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.446208] env[61768]: DEBUG nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Instance network_info: |[{"id": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "address": "fa:16:3e:b3:4f:18", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4f2467a-e3", "ovs_interfaceid": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 993.446621] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:4f:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4162774e-ec80-4d85-aeb4-fae77f197393', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 993.454097] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating folder: Project (e4c2ff63d9d94a41b0dbc7d597e762a8). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 993.454596] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3db6ef23-66c3-48b5-9b72-e2d73aef62f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.465955] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created folder: Project (e4c2ff63d9d94a41b0dbc7d597e762a8) in parent group-v265360. [ 993.466254] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating folder: Instances. Parent ref: group-v265528. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 993.466406] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89659e16-97c0-4bc3-8039-c763f9dd5e5b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.475682] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created folder: Instances in parent group-v265528. [ 993.475943] env[61768]: DEBUG oslo.service.loopingcall [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 993.476199] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 993.476442] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dae6f7cc-12e0-46b2-9094-4e391c77536f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.496414] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 993.496414] env[61768]: value = "task-1229356" [ 993.496414] env[61768]: _type = "Task" [ 993.496414] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.503734] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229356, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.550512] env[61768]: DEBUG oslo_vmware.api [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229353, 'name': PowerOffVM_Task, 'duration_secs': 0.160169} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.550854] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 993.550891] env[61768]: DEBUG nova.compute.manager [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.551686] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f65f9de8-178b-41eb-8c3a-a91bb6263fc2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.735753] env[61768]: DEBUG nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 993.893894] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 993.917278] env[61768]: DEBUG nova.scheduler.client.report [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.947741] env[61768]: DEBUG nova.network.neutron [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 994.006702] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229356, 'name': CreateVM_Task, 'duration_secs': 0.27872} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.006888] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 994.007557] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.007733] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.008073] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 994.008318] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd40c8a5-3e5b-4559-ac8c-ee937c0e6591 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.012662] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 994.012662] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5294773f-b3dc-94a0-a9ca-8d2a96ffc5b2" [ 994.012662] env[61768]: _type = "Task" [ 994.012662] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.019933] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5294773f-b3dc-94a0-a9ca-8d2a96ffc5b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.023344] env[61768]: DEBUG nova.network.neutron [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 994.063314] env[61768]: DEBUG oslo_concurrency.lockutils [None req-75abd078-8b68-4929-b925-65ebc97c6809 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.400188] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.400485] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.400741] env[61768]: INFO nova.compute.manager [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Attaching volume 31dc93bd-ffee-4257-90ed-c6adee08507a to /dev/sdb [ 994.423176] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.438769] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d69cb78-71c5-448b-9758-e3ba87b96c05 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.446467] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03196c7-2e9a-46e4-ac03-225b759dadf8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.449680] env[61768]: INFO nova.scheduler.client.report [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted allocations for instance 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f [ 994.463754] env[61768]: DEBUG nova.virt.block_device [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updating existing volume attachment record: 240f43e3-5dc0-4d8a-84d1-1a00d36fe1c9 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 994.522412] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5294773f-b3dc-94a0-a9ca-8d2a96ffc5b2, 'name': SearchDatastore_Task, 'duration_secs': 0.00954} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.522728] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.522969] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 994.523247] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.523450] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.523562] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 994.523828] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ec9f974e-41e9-4d0f-b640-0caa090530f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.525847] env[61768]: DEBUG oslo_concurrency.lockutils [req-6b358b4d-dae6-4ed0-87e7-7bf3351ed7ad req-382a3d18-454b-4be5-bf1a-8f6992d811ea service nova] Releasing lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 994.526379] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.526534] env[61768]: DEBUG nova.network.neutron [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 994.534671] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 994.535036] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 994.536076] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-111cedac-ad87-47eb-a4e6-133b66dc6f94 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.541518] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 994.541518] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523e5240-5637-b5d4-3def-a35dbeaa3a1d" [ 994.541518] env[61768]: _type = "Task" [ 994.541518] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.551310] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523e5240-5637-b5d4-3def-a35dbeaa3a1d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.745748] env[61768]: DEBUG nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 994.770794] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.771127] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.771377] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.771817] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.771960] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.772218] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.772789] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.772789] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.773067] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.773343] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.773636] env[61768]: DEBUG nova.virt.hardware [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.774720] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a448029e-02a1-451d-a8de-aef5b3a7efba {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.783956] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22972350-5658-43d0-a8aa-1abd2e6f0489 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.799102] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 994.805062] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Creating folder: Project (dcf2a3af66f54985b1021023e588c605). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 994.805456] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-076ce1c6-6c07-4405-b391-ce1156638e94 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.816230] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Created folder: Project (dcf2a3af66f54985b1021023e588c605) in parent group-v265360. [ 994.816442] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Creating folder: Instances. Parent ref: group-v265532. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 994.816738] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-946767b1-9ee6-4432-a99d-dc355b4b65f6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.825542] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Created folder: Instances in parent group-v265532. [ 994.825804] env[61768]: DEBUG oslo.service.loopingcall [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.826053] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 994.826296] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8bd78a6d-4c68-4873-92b7-3c77539c4a4a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.844692] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 994.844692] env[61768]: value = "task-1229360" [ 994.844692] env[61768]: _type = "Task" [ 994.844692] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.853561] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229360, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.957050] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9968829d-80c8-4bf0-bbcb-d07bed483a46 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.844s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.053018] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523e5240-5637-b5d4-3def-a35dbeaa3a1d, 'name': SearchDatastore_Task, 'duration_secs': 0.008782} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.053905] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cf600d15-51df-4132-aa86-299bd4fa821e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.060118] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 995.060118] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5206dc26-65d4-bc8e-6d87-12b72c668578" [ 995.060118] env[61768]: _type = "Task" [ 995.060118] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.065302] env[61768]: DEBUG nova.network.neutron [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 995.073499] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5206dc26-65d4-bc8e-6d87-12b72c668578, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.212182] env[61768]: DEBUG nova.network.neutron [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Updating instance_info_cache with network_info: [{"id": "5913b4b2-1038-4537-9cd9-b466dc48d869", "address": "fa:16:3e:89:a9:68", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5913b4b2-10", "ovs_interfaceid": "5913b4b2-1038-4537-9cd9-b466dc48d869", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.218172] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.219480] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.219480] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.219480] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.219480] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.220850] env[61768]: INFO nova.compute.manager [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Terminating instance [ 995.222618] env[61768]: DEBUG nova.compute.manager [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 995.222817] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 995.223652] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6c71f0e-8228-4cbf-add0-cd8a53975ba7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.231896] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 995.232142] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed89e25b-9e2c-4e25-b26a-e57669e907d7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.338266] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 995.338499] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 995.338691] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleting the datastore file [datastore1] e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.338961] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dcd13d9e-535f-49a0-938e-150e37088f16 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.349446] env[61768]: DEBUG oslo_vmware.api [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 995.349446] env[61768]: value = "task-1229362" [ 995.349446] env[61768]: _type = "Task" [ 995.349446] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.356872] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229360, 'name': CreateVM_Task, 'duration_secs': 0.302323} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.357393] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 995.358602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.358602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.358757] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 995.361784] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f13cb7e6-75da-4a12-a36e-e826391d7b49 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.363508] env[61768]: DEBUG oslo_vmware.api [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229362, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.366263] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 995.366263] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cc0cc2-e722-90e0-3bce-092195ea0abd" [ 995.366263] env[61768]: _type = "Task" [ 995.366263] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.376598] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cc0cc2-e722-90e0-3bce-092195ea0abd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.570954] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5206dc26-65d4-bc8e-6d87-12b72c668578, 'name': SearchDatastore_Task, 'duration_secs': 0.018842} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.571344] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.571742] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/e8095d6d-18c5-4ca9-8ac8-475aeeffc01a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 995.572055] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38883861-4cd8-46cf-922a-8c546908d535 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.579349] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 995.579349] env[61768]: value = "task-1229363" [ 995.579349] env[61768]: _type = "Task" [ 995.579349] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.586825] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.714802] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.715271] env[61768]: DEBUG nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Instance network_info: |[{"id": "5913b4b2-1038-4537-9cd9-b466dc48d869", "address": "fa:16:3e:89:a9:68", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5913b4b2-10", "ovs_interfaceid": "5913b4b2-1038-4537-9cd9-b466dc48d869", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 995.715717] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:a9:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4162774e-ec80-4d85-aeb4-fae77f197393', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5913b4b2-1038-4537-9cd9-b466dc48d869', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 995.724542] env[61768]: DEBUG oslo.service.loopingcall [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.724542] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 995.724711] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8e2f7f51-a2ed-4b1d-baea-9714e9ece6f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.746864] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 995.746864] env[61768]: value = "task-1229364" [ 995.746864] env[61768]: _type = "Task" [ 995.746864] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.757776] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229364, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.860767] env[61768]: DEBUG oslo_vmware.api [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229362, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135639} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.861165] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.861409] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 995.861650] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 995.861866] env[61768]: INFO nova.compute.manager [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Took 0.64 seconds to destroy the instance on the hypervisor. [ 995.862195] env[61768]: DEBUG oslo.service.loopingcall [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.862444] env[61768]: DEBUG nova.compute.manager [-] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.862542] env[61768]: DEBUG nova.network.neutron [-] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 995.879276] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52cc0cc2-e722-90e0-3bce-092195ea0abd, 'name': SearchDatastore_Task, 'duration_secs': 0.016379} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.879744] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.880219] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 995.880457] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.880684] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.880934] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 995.881356] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-32d9e365-90cd-47dc-8009-ee1e29826cf4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.896565] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 995.896808] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 995.897664] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-346d6dad-a039-4ff2-9d54-25f9ca816550 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.906753] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 995.906753] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529eab7d-98ec-2028-3859-05b5542edaf2" [ 995.906753] env[61768]: _type = "Task" [ 995.906753] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.918862] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529eab7d-98ec-2028-3859-05b5542edaf2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.097547] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.445925} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.098245] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/e8095d6d-18c5-4ca9-8ac8-475aeeffc01a.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 996.101719] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 996.102354] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3960645f-f702-43b2-90b7-a527e9f6f21e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.111848] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 996.111848] env[61768]: value = "task-1229365" [ 996.111848] env[61768]: _type = "Task" [ 996.111848] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.126125] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229365, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.149225] env[61768]: DEBUG nova.compute.manager [req-173ec95a-d53d-4c3a-89cc-54022bfdc6d7 req-7c663a86-4584-426e-8730-e44d98ab0043 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Received event network-vif-deleted-127c5490-87ca-460f-9891-b309eae19b78 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.149469] env[61768]: INFO nova.compute.manager [req-173ec95a-d53d-4c3a-89cc-54022bfdc6d7 req-7c663a86-4584-426e-8730-e44d98ab0043 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Neutron deleted interface 127c5490-87ca-460f-9891-b309eae19b78; detaching it from the instance and deleting it from the info cache [ 996.149522] env[61768]: DEBUG nova.network.neutron [req-173ec95a-d53d-4c3a-89cc-54022bfdc6d7 req-7c663a86-4584-426e-8730-e44d98ab0043 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.256860] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229364, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.304589] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "a96d4721-83c1-474f-8470-a34262e34fad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.304815] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.417341] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529eab7d-98ec-2028-3859-05b5542edaf2, 'name': SearchDatastore_Task, 'duration_secs': 0.059066} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.418133] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ce3a847-3341-40c3-8c7a-d5546a8fd436 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.423174] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 996.423174] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ccf152-31e7-f581-86f5-909b98b529cd" [ 996.423174] env[61768]: _type = "Task" [ 996.423174] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.430544] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ccf152-31e7-f581-86f5-909b98b529cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.623743] env[61768]: DEBUG nova.network.neutron [-] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.624915] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229365, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07046} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.625374] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 996.626176] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b7d1c44-102a-4804-b3e4-845a091a73e9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.647785] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/e8095d6d-18c5-4ca9-8ac8-475aeeffc01a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 996.648595] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e59762f4-fba5-4a10-b489-8193724275a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.664367] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c661729c-f1fa-448f-b935-4b6ecf9561ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.673878] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e0c13b-9b57-4d1e-829d-c4392d9d1e32 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.685464] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 996.685464] env[61768]: value = "task-1229367" [ 996.685464] env[61768]: _type = "Task" [ 996.685464] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.695439] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229367, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.703425] env[61768]: DEBUG nova.compute.manager [req-173ec95a-d53d-4c3a-89cc-54022bfdc6d7 req-7c663a86-4584-426e-8730-e44d98ab0043 service nova] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Detach interface failed, port_id=127c5490-87ca-460f-9891-b309eae19b78, reason: Instance e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 996.757230] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229364, 'name': CreateVM_Task, 'duration_secs': 0.588264} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.757230] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 996.757780] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.757959] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.758309] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 996.758569] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb9a4fed-40de-41e9-acee-4c15683f5122 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.763244] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 996.763244] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5286a286-6534-0168-be64-6f95d9eaab57" [ 996.763244] env[61768]: _type = "Task" [ 996.763244] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.770969] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5286a286-6534-0168-be64-6f95d9eaab57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.806984] env[61768]: DEBUG nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 996.933868] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ccf152-31e7-f581-86f5-909b98b529cd, 'name': SearchDatastore_Task, 'duration_secs': 0.105587} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 996.934170] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.934437] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f/5bd85959-cb58-4e63-ac1e-7a3ead11cc4f.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 996.934701] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-72333e7c-7bc4-480c-bf35-384b6dca345b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.945129] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 996.945129] env[61768]: value = "task-1229368" [ 996.945129] env[61768]: _type = "Task" [ 996.945129] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.953041] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229368, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.125869] env[61768]: INFO nova.compute.manager [-] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Took 1.26 seconds to deallocate network for instance. [ 997.196594] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229367, 'name': ReconfigVM_Task, 'duration_secs': 0.281129} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.196594] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/e8095d6d-18c5-4ca9-8ac8-475aeeffc01a.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 997.197318] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-eb27e211-537c-457e-8bd3-322aeb26dbe1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.203483] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 997.203483] env[61768]: value = "task-1229369" [ 997.203483] env[61768]: _type = "Task" [ 997.203483] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.211222] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229369, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.273763] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5286a286-6534-0168-be64-6f95d9eaab57, 'name': SearchDatastore_Task, 'duration_secs': 0.009348} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.274096] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.274350] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.274599] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.274754] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.274957] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.275294] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-af5b4d94-534d-4aec-8714-e81a4a465819 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.283614] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.283819] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 997.284579] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae3ea833-0ce7-45ac-b507-1f6c4937e430 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.289931] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 997.289931] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a4a5d-fcb1-0382-284d-227d980310da" [ 997.289931] env[61768]: _type = "Task" [ 997.289931] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.297571] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a4a5d-fcb1-0382-284d-227d980310da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.330076] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.330362] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.331866] env[61768]: INFO nova.compute.claims [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 997.455426] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229368, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.632545] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.712866] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229369, 'name': Rename_Task, 'duration_secs': 0.144037} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.713157] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 997.713407] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-794b3200-3baa-4973-9693-eff72a929d96 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.719451] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 997.719451] env[61768]: value = "task-1229370" [ 997.719451] env[61768]: _type = "Task" [ 997.719451] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.726601] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229370, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.799878] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a4a5d-fcb1-0382-284d-227d980310da, 'name': SearchDatastore_Task, 'duration_secs': 0.008604} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.800669] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60091840-4948-48ac-a8d6-d3cfd9f10203 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.805493] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 997.805493] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ce9c69-6357-92c8-59c1-68f34a7ee997" [ 997.805493] env[61768]: _type = "Task" [ 997.805493] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.812706] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ce9c69-6357-92c8-59c1-68f34a7ee997, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.957317] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229368, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705449} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.957650] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f/5bd85959-cb58-4e63-ac1e-7a3ead11cc4f.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 997.957947] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 997.958256] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c024f21e-8254-4e57-b9ea-6b72209f0b38 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.964558] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 997.964558] env[61768]: value = "task-1229371" [ 997.964558] env[61768]: _type = "Task" [ 997.964558] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.972523] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229371, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.230628] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229370, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.316824] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ce9c69-6357-92c8-59c1-68f34a7ee997, 'name': SearchDatastore_Task, 'duration_secs': 0.008546} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.317047] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.317331] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/aa568a72-afba-4a25-a048-ca62191f9212.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 998.317662] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e04b8320-fc4b-43cf-aa5e-183494e6753d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.324788] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 998.324788] env[61768]: value = "task-1229372" [ 998.324788] env[61768]: _type = "Task" [ 998.324788] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.332805] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229372, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.458733] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bc530fe-18c4-4442-be8e-21b0c749e640 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.469133] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f48390f-2629-4295-8228-dee271da0507 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.476801] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229371, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070225} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.501035] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 998.502023] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939d09ab-b089-444f-8170-0171498c4a1a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.504835] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e03e043-9d1b-45be-9923-c604ed848877 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.517890] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d748b2-f5f8-47b4-8f96-91b02807b367 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.529457] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f/5bd85959-cb58-4e63-ac1e-7a3ead11cc4f.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 998.529740] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53a9f7e4-0d0b-4b89-95b1-8324f48159c4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.555633] env[61768]: DEBUG nova.compute.provider_tree [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.558304] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 998.558304] env[61768]: value = "task-1229373" [ 998.558304] env[61768]: _type = "Task" [ 998.558304] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.566793] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229373, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.732910] env[61768]: DEBUG oslo_vmware.api [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229370, 'name': PowerOnVM_Task, 'duration_secs': 0.632744} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.733303] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 998.733431] env[61768]: INFO nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Took 8.44 seconds to spawn the instance on the hypervisor. [ 998.733611] env[61768]: DEBUG nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.734423] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76aaa3ab-7d27-46ca-9362-123d2c4ef383 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.835696] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229372, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.510249} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.836040] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/aa568a72-afba-4a25-a048-ca62191f9212.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 998.836397] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 998.836767] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7780ee7c-75b9-4bbb-9400-d8a696dbcd57 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.843601] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 998.843601] env[61768]: value = "task-1229374" [ 998.843601] env[61768]: _type = "Task" [ 998.843601] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.851151] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229374, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.009580] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 999.009580] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265531', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'name': 'volume-31dc93bd-ffee-4257-90ed-c6adee08507a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'serial': '31dc93bd-ffee-4257-90ed-c6adee08507a'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 999.010436] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7458edca-0d31-4e9b-b69a-d7b15b1faf39 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.027304] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-317451aa-97e6-4a8a-ae7f-75e52cfb5d3e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.054650] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] volume-31dc93bd-ffee-4257-90ed-c6adee08507a/volume-31dc93bd-ffee-4257-90ed-c6adee08507a.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.054650] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-557589da-07c9-4348-b16a-67de2c0d68a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.067497] env[61768]: DEBUG nova.scheduler.client.report [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 999.080409] env[61768]: DEBUG oslo_vmware.api [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 999.080409] env[61768]: value = "task-1229375" [ 999.080409] env[61768]: _type = "Task" [ 999.080409] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.087211] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229373, 'name': ReconfigVM_Task, 'duration_secs': 0.35013} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.087837] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Reconfigured VM instance instance-0000006b to attach disk [datastore1] 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f/5bd85959-cb58-4e63-ac1e-7a3ead11cc4f.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.088510] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69a4e9c0-29e5-4656-a6d7-c926010b74f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.093706] env[61768]: DEBUG oslo_vmware.api [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229375, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.098501] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 999.098501] env[61768]: value = "task-1229376" [ 999.098501] env[61768]: _type = "Task" [ 999.098501] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.107611] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229376, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.254682] env[61768]: INFO nova.compute.manager [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Took 13.21 seconds to build instance. [ 999.353781] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229374, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062387} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.354077] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.354817] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71dd1740-5514-4f25-8631-55e5bd45492a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.378352] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/aa568a72-afba-4a25-a048-ca62191f9212.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.378615] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eeaabab5-b7b6-40cf-b9ab-981ad5816716 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.398070] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 999.398070] env[61768]: value = "task-1229377" [ 999.398070] env[61768]: _type = "Task" [ 999.398070] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.407167] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229377, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.576082] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.246s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.576632] env[61768]: DEBUG nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 999.579276] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.947s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.579509] env[61768]: DEBUG nova.objects.instance [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'resources' on Instance uuid e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 999.589722] env[61768]: DEBUG oslo_vmware.api [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229375, 'name': ReconfigVM_Task, 'duration_secs': 0.369534} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.590493] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfigured VM instance instance-00000061 to attach disk [datastore1] volume-31dc93bd-ffee-4257-90ed-c6adee08507a/volume-31dc93bd-ffee-4257-90ed-c6adee08507a.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.595463] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-421c6de8-6711-4d6e-81c3-36978dc4f193 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.615015] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229376, 'name': Rename_Task, 'duration_secs': 0.177609} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.616258] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 999.616597] env[61768]: DEBUG oslo_vmware.api [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 999.616597] env[61768]: value = "task-1229378" [ 999.616597] env[61768]: _type = "Task" [ 999.616597] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.616812] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49031abd-66eb-4761-a670-3023bea70293 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.627194] env[61768]: DEBUG oslo_vmware.api [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229378, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.628409] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 999.628409] env[61768]: value = "task-1229379" [ 999.628409] env[61768]: _type = "Task" [ 999.628409] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.636761] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229379, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.757397] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d4007545-f670-466f-993d-4bbdb6ebf656 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.717s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.908873] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229377, 'name': ReconfigVM_Task, 'duration_secs': 0.28516} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.909194] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Reconfigured VM instance instance-0000006a to attach disk [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/aa568a72-afba-4a25-a048-ca62191f9212.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 999.909798] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5cc9fa0-fb00-4dbf-942d-a7d8be6565a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.917349] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 999.917349] env[61768]: value = "task-1229380" [ 999.917349] env[61768]: _type = "Task" [ 999.917349] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.926256] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229380, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.084173] env[61768]: DEBUG nova.compute.utils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1000.088561] env[61768]: DEBUG nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1000.088561] env[61768]: DEBUG nova.network.neutron [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1000.130099] env[61768]: DEBUG oslo_vmware.api [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229378, 'name': ReconfigVM_Task, 'duration_secs': 0.147914} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.133312] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265531', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'name': 'volume-31dc93bd-ffee-4257-90ed-c6adee08507a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'serial': '31dc93bd-ffee-4257-90ed-c6adee08507a'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1000.138134] env[61768]: DEBUG nova.policy [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'efd1aa7c174648888b58f6a6df6f82a5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dfce0cf4a5d14042a929dac65aeb2ae3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 1000.143559] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229379, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.204530] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8506dc2-e5f5-4937-9981-e5903f99c511 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.213370] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-847a2fdd-0dbc-4a5c-973b-7cd48f43aec8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.246626] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e2ab993-003e-47d0-8f0f-1f46ebcf656d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.255209] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d64b484-124f-4fbb-bd7b-2b293fb62541 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.268800] env[61768]: DEBUG nova.compute.provider_tree [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.400976] env[61768]: DEBUG nova.network.neutron [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Successfully created port: 3c8e6155-6935-4025-a622-f7e4f08a0ec3 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1000.431742] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229380, 'name': Rename_Task, 'duration_secs': 0.156166} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.431899] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1000.432358] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f60191b0-495d-4b70-8ed5-f7794f16604a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.440414] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1000.440414] env[61768]: value = "task-1229381" [ 1000.440414] env[61768]: _type = "Task" [ 1000.440414] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.450676] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229381, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.589421] env[61768]: DEBUG nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.639869] env[61768]: DEBUG oslo_vmware.api [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229379, 'name': PowerOnVM_Task, 'duration_secs': 0.632979} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.642156] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1000.642376] env[61768]: INFO nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Took 5.90 seconds to spawn the instance on the hypervisor. [ 1000.642563] env[61768]: DEBUG nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1000.643756] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfdd49c5-e538-414d-8ed4-9bb234d5f543 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.772022] env[61768]: DEBUG nova.scheduler.client.report [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.952287] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229381, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.160422] env[61768]: INFO nova.compute.manager [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Took 13.65 seconds to build instance. [ 1001.186602] env[61768]: DEBUG nova.objects.instance [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid e1e90c98-5a65-473c-b508-b45fd93e31ad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.276633] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.302742] env[61768]: INFO nova.scheduler.client.report [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted allocations for instance e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa [ 1001.450966] env[61768]: DEBUG oslo_vmware.api [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229381, 'name': PowerOnVM_Task, 'duration_secs': 0.645184} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.451293] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.451508] env[61768]: INFO nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Took 8.92 seconds to spawn the instance on the hypervisor. [ 1001.451729] env[61768]: DEBUG nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.452526] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe45371-6440-486f-8b0d-16c9d71e6af4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.599092] env[61768]: DEBUG nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1001.620670] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.620955] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.621207] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.621490] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.621608] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.621772] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.622025] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.622245] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.622414] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.622594] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.622778] env[61768]: DEBUG nova.virt.hardware [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.623658] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5b3bfbf-fdb6-4f4f-acd5-fb6356aa342c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.632350] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9def0068-e2cf-4a59-b19f-2d5da397e8cc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.662364] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9ed7b8d8-4073-41ca-87f3-7bcbb3250edc tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.164s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.692410] env[61768]: DEBUG oslo_concurrency.lockutils [None req-b575dce6-2ec4-43dc-9a98-22bd67a5540c tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.292s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.745958] env[61768]: DEBUG nova.compute.manager [None req-9839427b-97bf-4cca-848c-216c52cefaa1 tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1001.746913] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efcf5e60-3c4d-4df5-9a46-4daff5b26f89 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.812502] env[61768]: DEBUG oslo_concurrency.lockutils [None req-24864d88-14e3-4efe-a79b-8dbc93306cfa tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.594s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.824402] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.824678] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.824897] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.825103] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.825285] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.829761] env[61768]: INFO nova.compute.manager [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Terminating instance [ 1001.831075] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "refresh_cache-5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.831075] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquired lock "refresh_cache-5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.831075] env[61768]: DEBUG nova.network.neutron [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1001.894019] env[61768]: DEBUG nova.compute.manager [req-29edd729-de92-4826-b7b0-8f3b390d4a4b req-daff6b57-d0db-4415-abb7-1f7936e39a93 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Received event network-vif-plugged-3c8e6155-6935-4025-a622-f7e4f08a0ec3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1001.894342] env[61768]: DEBUG oslo_concurrency.lockutils [req-29edd729-de92-4826-b7b0-8f3b390d4a4b req-daff6b57-d0db-4415-abb7-1f7936e39a93 service nova] Acquiring lock "a96d4721-83c1-474f-8470-a34262e34fad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1001.894546] env[61768]: DEBUG oslo_concurrency.lockutils [req-29edd729-de92-4826-b7b0-8f3b390d4a4b req-daff6b57-d0db-4415-abb7-1f7936e39a93 service nova] Lock "a96d4721-83c1-474f-8470-a34262e34fad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.894734] env[61768]: DEBUG oslo_concurrency.lockutils [req-29edd729-de92-4826-b7b0-8f3b390d4a4b req-daff6b57-d0db-4415-abb7-1f7936e39a93 service nova] Lock "a96d4721-83c1-474f-8470-a34262e34fad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.894909] env[61768]: DEBUG nova.compute.manager [req-29edd729-de92-4826-b7b0-8f3b390d4a4b req-daff6b57-d0db-4415-abb7-1f7936e39a93 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] No waiting events found dispatching network-vif-plugged-3c8e6155-6935-4025-a622-f7e4f08a0ec3 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1001.895138] env[61768]: WARNING nova.compute.manager [req-29edd729-de92-4826-b7b0-8f3b390d4a4b req-daff6b57-d0db-4415-abb7-1f7936e39a93 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Received unexpected event network-vif-plugged-3c8e6155-6935-4025-a622-f7e4f08a0ec3 for instance with vm_state building and task_state spawning. [ 1001.968864] env[61768]: INFO nova.compute.manager [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Took 15.28 seconds to build instance. [ 1001.987901] env[61768]: DEBUG nova.network.neutron [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Successfully updated port: 3c8e6155-6935-4025-a622-f7e4f08a0ec3 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1002.258034] env[61768]: INFO nova.compute.manager [None req-9839427b-97bf-4cca-848c-216c52cefaa1 tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] instance snapshotting [ 1002.258518] env[61768]: DEBUG nova.objects.instance [None req-9839427b-97bf-4cca-848c-216c52cefaa1 tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lazy-loading 'flavor' on Instance uuid 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1002.350376] env[61768]: DEBUG nova.network.neutron [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1002.390266] env[61768]: INFO nova.compute.manager [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Rescuing [ 1002.390563] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.390722] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.390897] env[61768]: DEBUG nova.network.neutron [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.471654] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8e722c6-801d-4675-9627-f699c1769244 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "aa568a72-afba-4a25-a048-ca62191f9212" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.788s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.493103] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.493382] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.493660] env[61768]: DEBUG nova.network.neutron [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.564431] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1002.564678] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.606804] env[61768]: DEBUG nova.network.neutron [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1002.629220] env[61768]: DEBUG nova.network.neutron [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.765707] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1667582b-2afe-4f82-8296-9b534ddb7b19 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.784375] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a720a17-c2b2-491c-a3eb-5dd5f26ea4bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.801401] env[61768]: DEBUG nova.network.neutron [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance_info_cache with network_info: [{"id": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "address": "fa:16:3e:b9:94:f5", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8e6155-69", "ovs_interfaceid": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.068273] env[61768]: DEBUG nova.compute.utils [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1003.125096] env[61768]: DEBUG nova.network.neutron [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Updating instance_info_cache with network_info: [{"id": "5913b4b2-1038-4537-9cd9-b466dc48d869", "address": "fa:16:3e:89:a9:68", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5913b4b2-10", "ovs_interfaceid": "5913b4b2-1038-4537-9cd9-b466dc48d869", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1003.132148] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Releasing lock "refresh_cache-5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.132583] env[61768]: DEBUG nova.compute.manager [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.132786] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.133722] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef92bbb-0c81-4f33-9a5c-75936f0f11d1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.144238] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.144477] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6f0c062-9f67-4e0f-93cb-5a8d6e59ac2c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.151277] env[61768]: DEBUG oslo_vmware.api [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 1003.151277] env[61768]: value = "task-1229382" [ 1003.151277] env[61768]: _type = "Task" [ 1003.151277] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.154927] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.155225] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.155384] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.155577] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.155747] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.157997] env[61768]: INFO nova.compute.manager [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Terminating instance [ 1003.162412] env[61768]: DEBUG oslo_vmware.api [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229382, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.162997] env[61768]: DEBUG nova.compute.manager [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1003.163207] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1003.163970] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be3ec9b-edf3-4959-b819-1705e97ad648 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.171361] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1003.171555] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e743fbc-37b3-4392-bc71-b5693c552ece {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.177869] env[61768]: DEBUG oslo_vmware.api [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 1003.177869] env[61768]: value = "task-1229383" [ 1003.177869] env[61768]: _type = "Task" [ 1003.177869] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.185764] env[61768]: DEBUG oslo_vmware.api [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229383, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.298044] env[61768]: DEBUG nova.compute.manager [None req-9839427b-97bf-4cca-848c-216c52cefaa1 tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Instance disappeared during snapshot {{(pid=61768) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 1003.304032] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.304336] env[61768]: DEBUG nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Instance network_info: |[{"id": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "address": "fa:16:3e:b9:94:f5", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8e6155-69", "ovs_interfaceid": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1003.304777] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b9:94:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '98e21102-8954-4f6f-b1e6-5d764a53aa22', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c8e6155-6935-4025-a622-f7e4f08a0ec3', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1003.312991] env[61768]: DEBUG oslo.service.loopingcall [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.313299] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1003.313601] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-00c998e6-f907-49e4-acac-c3de8db330c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.336397] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1003.336397] env[61768]: value = "task-1229384" [ 1003.336397] env[61768]: _type = "Task" [ 1003.336397] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.348668] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229384, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.451325] env[61768]: DEBUG nova.compute.manager [None req-9839427b-97bf-4cca-848c-216c52cefaa1 tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Found 0 images (rotation: 2) {{(pid=61768) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1003.570975] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.628205] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-aa568a72-afba-4a25-a048-ca62191f9212" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.661282] env[61768]: DEBUG oslo_vmware.api [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229382, 'name': PowerOffVM_Task, 'duration_secs': 0.145155} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.662229] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1003.662381] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1003.662660] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-918c0fce-1e07-43ea-91d7-6eef599eb4a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.688498] env[61768]: DEBUG oslo_vmware.api [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229383, 'name': PowerOffVM_Task, 'duration_secs': 0.213234} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.688706] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1003.688883] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1003.689173] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d605f4c7-fc46-49bc-b479-238877440cdf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.702635] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1003.702858] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1003.703081] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Deleting the datastore file [datastore1] 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.703378] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7b02f25-3f5e-46e3-b9d6-781a0be1609f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.711298] env[61768]: DEBUG oslo_vmware.api [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for the task: (returnval){ [ 1003.711298] env[61768]: value = "task-1229387" [ 1003.711298] env[61768]: _type = "Task" [ 1003.711298] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.720162] env[61768]: DEBUG oslo_vmware.api [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229387, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.850666] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229384, 'name': CreateVM_Task, 'duration_secs': 0.482815} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.850870] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1003.851649] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.851831] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.852199] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.852476] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bea7ebc-df03-4075-9502-23bc63fb55fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.858376] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1003.858376] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c02f84-1503-a604-ea34-7fc0d4c98c08" [ 1003.858376] env[61768]: _type = "Task" [ 1003.858376] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.867860] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c02f84-1503-a604-ea34-7fc0d4c98c08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.961865] env[61768]: DEBUG nova.compute.manager [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Received event network-changed-3c8e6155-6935-4025-a622-f7e4f08a0ec3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.962032] env[61768]: DEBUG nova.compute.manager [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Refreshing instance network info cache due to event network-changed-3c8e6155-6935-4025-a622-f7e4f08a0ec3. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1003.962268] env[61768]: DEBUG oslo_concurrency.lockutils [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] Acquiring lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.962462] env[61768]: DEBUG oslo_concurrency.lockutils [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] Acquired lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.962653] env[61768]: DEBUG nova.network.neutron [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Refreshing network info cache for port 3c8e6155-6935-4025-a622-f7e4f08a0ec3 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1003.994821] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1003.995084] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1003.995295] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleting the datastore file [datastore1] 2c85de5d-0111-40a6-a8bc-69c7eba0393d {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1003.995611] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26dde152-ec96-4c8f-a6aa-0089b5b2f577 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.002828] env[61768]: DEBUG oslo_vmware.api [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for the task: (returnval){ [ 1004.002828] env[61768]: value = "task-1229388" [ 1004.002828] env[61768]: _type = "Task" [ 1004.002828] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.011873] env[61768]: DEBUG oslo_vmware.api [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229388, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.159602] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.159953] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2c2f11c1-4140-48b2-abe9-ec3011595972 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.168305] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1004.168305] env[61768]: value = "task-1229389" [ 1004.168305] env[61768]: _type = "Task" [ 1004.168305] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.177927] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229389, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.223295] env[61768]: DEBUG oslo_vmware.api [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Task: {'id': task-1229387, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157425} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.223552] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.223765] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1004.223953] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1004.224153] env[61768]: INFO nova.compute.manager [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Took 1.09 seconds to destroy the instance on the hypervisor. [ 1004.224409] env[61768]: DEBUG oslo.service.loopingcall [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.224617] env[61768]: DEBUG nova.compute.manager [-] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.224712] env[61768]: DEBUG nova.network.neutron [-] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.246165] env[61768]: DEBUG nova.network.neutron [-] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1004.369529] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52c02f84-1503-a604-ea34-7fc0d4c98c08, 'name': SearchDatastore_Task, 'duration_secs': 0.023237} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.369851] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.370111] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.370363] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.370520] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.370706] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.370975] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-300c3c11-cc48-4326-a0c1-abfd6c009f60 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.388882] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.389122] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1004.389899] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f067a990-560a-4b50-969f-737f93068d44 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.396886] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1004.396886] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52347710-2ef0-0f34-c7c5-b9300956e583" [ 1004.396886] env[61768]: _type = "Task" [ 1004.396886] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.405477] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52347710-2ef0-0f34-c7c5-b9300956e583, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.512752] env[61768]: DEBUG oslo_vmware.api [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Task: {'id': task-1229388, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266504} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.513388] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1004.513610] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1004.513805] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1004.513987] env[61768]: INFO nova.compute.manager [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Took 1.35 seconds to destroy the instance on the hypervisor. [ 1004.514255] env[61768]: DEBUG oslo.service.loopingcall [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.514458] env[61768]: DEBUG nova.compute.manager [-] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1004.514554] env[61768]: DEBUG nova.network.neutron [-] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1004.629817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.630187] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.630471] env[61768]: INFO nova.compute.manager [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Attaching volume da44c25e-cb5b-4c4c-94b3-b0af5fd7a004 to /dev/sdc [ 1004.662538] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88d10477-e268-473e-9fe1-cf080210c430 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.672947] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-265ae518-b806-4c6a-a3e5-98be1c28948a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.680131] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229389, 'name': PowerOffVM_Task, 'duration_secs': 0.265695} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.680395] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1004.681138] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd996176-ef0d-43c9-8e7a-98031c6eb2ed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.687876] env[61768]: DEBUG nova.virt.block_device [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updating existing volume attachment record: 071d246f-f9d8-467c-b289-f381565c00d1 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1004.704982] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd5fbc85-2482-4be9-88bd-65a890c310ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.749940] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1004.750278] env[61768]: DEBUG nova.network.neutron [-] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.751407] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93805493-fa4e-4bd3-9fc6-18e570c60855 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.762033] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1004.762033] env[61768]: value = "task-1229390" [ 1004.762033] env[61768]: _type = "Task" [ 1004.762033] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.771130] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1004.771130] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.771753] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.771753] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.771753] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.772826] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e9fe3082-a242-4a38-b13a-f36f556189af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.776085] env[61768]: DEBUG nova.network.neutron [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updated VIF entry in instance network info cache for port 3c8e6155-6935-4025-a622-f7e4f08a0ec3. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1004.776380] env[61768]: DEBUG nova.network.neutron [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance_info_cache with network_info: [{"id": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "address": "fa:16:3e:b9:94:f5", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8e6155-69", "ovs_interfaceid": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.781534] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.781650] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1004.782450] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83be98b7-eb9b-4cf5-8b0c-b93328a9a61c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.788077] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1004.788077] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527f506a-7c24-90a4-6fe6-b445617f4ede" [ 1004.788077] env[61768]: _type = "Task" [ 1004.788077] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.797599] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527f506a-7c24-90a4-6fe6-b445617f4ede, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.907411] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52347710-2ef0-0f34-c7c5-b9300956e583, 'name': SearchDatastore_Task, 'duration_secs': 0.012607} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.908281] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b9e8f98-15ff-45a9-a929-9dfeeddc7d5e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.913870] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1004.913870] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5211fad0-0691-6306-8ac9-87571576a52f" [ 1004.913870] env[61768]: _type = "Task" [ 1004.913870] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.921308] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5211fad0-0691-6306-8ac9-87571576a52f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.254639] env[61768]: INFO nova.compute.manager [-] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Took 1.03 seconds to deallocate network for instance. [ 1005.278896] env[61768]: DEBUG oslo_concurrency.lockutils [req-4e38362b-ccf2-4a40-abf7-6225836afa95 req-5d58f2f1-6499-44e0-9bf3-70ba386e5b80 service nova] Releasing lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.300236] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]527f506a-7c24-90a4-6fe6-b445617f4ede, 'name': SearchDatastore_Task, 'duration_secs': 0.02543} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.301130] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d00b7c11-5b3a-4322-8ef2-1a7c28dafd71 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.307406] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1005.307406] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a2efa-9082-2534-5205-4131bdbcc44f" [ 1005.307406] env[61768]: _type = "Task" [ 1005.307406] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.317682] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a2efa-9082-2534-5205-4131bdbcc44f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.351795] env[61768]: DEBUG nova.network.neutron [-] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.425873] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5211fad0-0691-6306-8ac9-87571576a52f, 'name': SearchDatastore_Task, 'duration_secs': 0.025144} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.426270] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.426700] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] a96d4721-83c1-474f-8470-a34262e34fad/a96d4721-83c1-474f-8470-a34262e34fad.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1005.426997] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9a276b7d-7dbf-4c48-bac3-2dabe279f70a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.437056] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1005.437056] env[61768]: value = "task-1229392" [ 1005.437056] env[61768]: _type = "Task" [ 1005.437056] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.445987] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.761052] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.761420] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.761668] env[61768]: DEBUG nova.objects.instance [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lazy-loading 'resources' on Instance uuid 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.824315] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]528a2efa-9082-2534-5205-4131bdbcc44f, 'name': SearchDatastore_Task, 'duration_secs': 0.011764} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.824634] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.824915] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. {{(pid=61768) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1005.825300] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1a1e32e-e6ce-41d9-8159-91f6907664c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.835916] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1005.835916] env[61768]: value = "task-1229393" [ 1005.835916] env[61768]: _type = "Task" [ 1005.835916] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.847533] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.854257] env[61768]: INFO nova.compute.manager [-] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Took 1.34 seconds to deallocate network for instance. [ 1005.947727] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435718} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.947927] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] a96d4721-83c1-474f-8470-a34262e34fad/a96d4721-83c1-474f-8470-a34262e34fad.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1005.948176] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.948467] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-230fb684-4d96-48c8-a473-c161f48bfbd4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.956631] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1005.956631] env[61768]: value = "task-1229394" [ 1005.956631] env[61768]: _type = "Task" [ 1005.956631] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.966378] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.990359] env[61768]: DEBUG nova.compute.manager [req-3921e705-a558-4bab-b975-4b43052add8e req-296e837d-ffb1-42be-974d-109218fdd132 service nova] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Received event network-vif-deleted-49a82793-2751-49db-b262-593b386ea68d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.081422] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.081761] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.082027] env[61768]: INFO nova.compute.manager [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Shelving [ 1006.348300] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229393, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.469776} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.350753] env[61768]: INFO nova.virt.vmwareapi.ds_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. [ 1006.351730] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342a42a2-72e9-413b-8f7c-302d7d5e368f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.370174] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.377526] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.379906] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b622b42b-93d5-49df-bdd0-16bb176984f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.399221] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1006.399221] env[61768]: value = "task-1229395" [ 1006.399221] env[61768]: _type = "Task" [ 1006.399221] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.409352] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229395, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.415128] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ec8252-c9e3-4ea2-ba19-14ffef3bc7af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.422398] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-303418b3-3f40-4042-a17b-71b3c83f73a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.453692] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370b8d89-69ca-42df-8089-b4ae705ace4c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.463638] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a07cd268-ec78-4aa6-b84f-200f438b9561 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.470045] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084787} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.470712] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1006.471621] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debafcae-02bc-429f-a775-69cbbc8bd752 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.483330] env[61768]: DEBUG nova.compute.provider_tree [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.503645] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] a96d4721-83c1-474f-8470-a34262e34fad/a96d4721-83c1-474f-8470-a34262e34fad.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.504413] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b04a278d-894a-49e7-9c17-78e41273b3fe {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.526632] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1006.526632] env[61768]: value = "task-1229396" [ 1006.526632] env[61768]: _type = "Task" [ 1006.526632] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.536456] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229396, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.590015] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1006.590308] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fa4af39a-7ded-4428-a9f5-49965ca9fb90 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.597897] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1006.597897] env[61768]: value = "task-1229397" [ 1006.597897] env[61768]: _type = "Task" [ 1006.597897] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.606525] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229397, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.909548] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229395, 'name': ReconfigVM_Task, 'duration_secs': 0.355992} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.909924] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Reconfigured VM instance instance-0000006a to attach disk [datastore1] aa568a72-afba-4a25-a048-ca62191f9212/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.910772] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b54ace-6555-4bb8-9424-f80201768c98 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.939068] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af4cae64-73f3-410b-aae1-65bea86dafa2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.953582] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1006.953582] env[61768]: value = "task-1229398" [ 1006.953582] env[61768]: _type = "Task" [ 1006.953582] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.962906] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229398, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.984307] env[61768]: DEBUG nova.scheduler.client.report [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1007.036577] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229396, 'name': ReconfigVM_Task, 'duration_secs': 0.290679} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.036886] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Reconfigured VM instance instance-0000006c to attach disk [datastore2] a96d4721-83c1-474f-8470-a34262e34fad/a96d4721-83c1-474f-8470-a34262e34fad.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.037596] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4909d51a-00e5-4dd8-8a17-6dfae4a8a8d6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.044544] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1007.044544] env[61768]: value = "task-1229400" [ 1007.044544] env[61768]: _type = "Task" [ 1007.044544] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.053125] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229400, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.109605] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229397, 'name': PowerOffVM_Task, 'duration_secs': 0.188585} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.109907] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1007.110745] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d489ae0-25de-45aa-bfe4-6013c02fa80a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.130302] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fb9c43e-a4de-40c4-8e7b-6bbac27aba09 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.463636] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229398, 'name': ReconfigVM_Task, 'duration_secs': 0.188053} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.464052] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1007.464196] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32668013-e0a3-40d8-9bc3-14a0829ac39a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.470585] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1007.470585] env[61768]: value = "task-1229401" [ 1007.470585] env[61768]: _type = "Task" [ 1007.470585] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.478020] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229401, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.488844] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.727s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.490995] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.121s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.491323] env[61768]: DEBUG nova.objects.instance [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lazy-loading 'resources' on Instance uuid 2c85de5d-0111-40a6-a8bc-69c7eba0393d {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1007.512257] env[61768]: INFO nova.scheduler.client.report [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Deleted allocations for instance 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f [ 1007.555721] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229400, 'name': Rename_Task, 'duration_secs': 0.149278} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.556703] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1007.556972] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afc502d5-0a86-42ee-bb49-6644c3c826e1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.565035] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1007.565035] env[61768]: value = "task-1229402" [ 1007.565035] env[61768]: _type = "Task" [ 1007.565035] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.573215] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229402, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.642594] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Creating Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1007.642594] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1ab8c7df-111c-4773-aaa8-2c419df4d863 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.653505] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1007.653505] env[61768]: value = "task-1229403" [ 1007.653505] env[61768]: _type = "Task" [ 1007.653505] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.662432] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229403, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.992719] env[61768]: DEBUG oslo_vmware.api [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229401, 'name': PowerOnVM_Task, 'duration_secs': 0.487958} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.995644] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1007.998351] env[61768]: DEBUG nova.compute.manager [None req-e1e4862c-f903-4147-a1ca-7e4320c10bf8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.999297] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d524279a-e4a4-4ee7-8ca4-3d368375cad6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.019588] env[61768]: DEBUG oslo_concurrency.lockutils [None req-ea1d4f05-8767-4042-833c-45f163a5e76a tempest-ServersAaction247Test-267783746 tempest-ServersAaction247Test-267783746-project-member] Lock "5bd85959-cb58-4e63-ac1e-7a3ead11cc4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.195s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.075662] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229402, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.100261] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c311bff4-d855-416e-aaa1-ca22aa11b032 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.108671] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8c754a-1cd8-4560-b177-dc73df4711d5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.139754] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f12abb2c-29ec-4fb3-b2b3-6d511b6add61 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.148174] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7cdb341-3b29-4d7e-b6f4-5e44ffc02637 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.162604] env[61768]: DEBUG nova.compute.provider_tree [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.172565] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229403, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.577715] env[61768]: DEBUG oslo_vmware.api [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229402, 'name': PowerOnVM_Task, 'duration_secs': 0.681144} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.578091] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1008.578405] env[61768]: INFO nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Took 6.98 seconds to spawn the instance on the hypervisor. [ 1008.578670] env[61768]: DEBUG nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.579483] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032762b2-781b-462c-9b51-d72e30f25a14 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.671997] env[61768]: DEBUG nova.scheduler.client.report [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.675118] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229403, 'name': CreateSnapshot_Task, 'duration_secs': 0.730304} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.675562] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Created Snapshot of the VM instance {{(pid=61768) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1008.676387] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9f5c223-e425-49a4-9bb1-0c07f5d69d73 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.097775] env[61768]: INFO nova.compute.manager [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Took 11.79 seconds to build instance. [ 1009.177622] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.195875] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Creating linked-clone VM from snapshot {{(pid=61768) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1009.196316] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fa25095d-28a3-447f-a5bb-aef82ac1a56d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.201721] env[61768]: INFO nova.scheduler.client.report [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Deleted allocations for instance 2c85de5d-0111-40a6-a8bc-69c7eba0393d [ 1009.211026] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1009.211026] env[61768]: value = "task-1229404" [ 1009.211026] env[61768]: _type = "Task" [ 1009.211026] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.219250] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229404, 'name': CloneVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.264549] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1009.264820] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265537', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'name': 'volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'serial': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1009.265732] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c350dcc5-7d2e-4307-be13-f4826985ff2e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.283748] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c741d48-6bc6-4b5b-8991-a61693be9435 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.315507] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004/volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1009.315928] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17036a52-5bc2-4243-af59-85162aa86d2a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.339351] env[61768]: DEBUG oslo_vmware.api [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1009.339351] env[61768]: value = "task-1229405" [ 1009.339351] env[61768]: _type = "Task" [ 1009.339351] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.350032] env[61768]: DEBUG oslo_vmware.api [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229405, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.433878] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.434154] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.600224] env[61768]: DEBUG oslo_concurrency.lockutils [None req-7e07954c-bb37-4f2a-a436-689b9f6ee4b3 tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.295s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.716688] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e0cc6fa0-912a-4250-866a-7247d3eaae54 tempest-ServersTestJSON-949727068 tempest-ServersTestJSON-949727068-project-member] Lock "2c85de5d-0111-40a6-a8bc-69c7eba0393d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.561s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.723906] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229404, 'name': CloneVM_Task} progress is 94%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.849784] env[61768]: DEBUG oslo_vmware.api [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229405, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.937064] env[61768]: DEBUG nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.225080] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229404, 'name': CloneVM_Task} progress is 95%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.350870] env[61768]: DEBUG oslo_vmware.api [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229405, 'name': ReconfigVM_Task, 'duration_secs': 0.597873} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.351205] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfigured VM instance instance-00000061 to attach disk [datastore2] volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004/volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1010.356210] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7551d7b-cea2-41f4-9e9d-ce48ef7b244f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.374016] env[61768]: DEBUG oslo_vmware.api [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1010.374016] env[61768]: value = "task-1229406" [ 1010.374016] env[61768]: _type = "Task" [ 1010.374016] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.383932] env[61768]: DEBUG oslo_vmware.api [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229406, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.465186] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.465623] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.469196] env[61768]: INFO nova.compute.claims [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1010.479354] env[61768]: DEBUG nova.compute.manager [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Stashing vm_state: active {{(pid=61768) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1010.724463] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229404, 'name': CloneVM_Task, 'duration_secs': 1.138315} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.724843] env[61768]: INFO nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Created linked-clone VM from snapshot [ 1010.725495] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac7d296-9544-412e-b680-8e410e87785b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.735593] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Uploading image 9b2e869d-9e9c-4311-aae2-355101bc801a {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1010.767859] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1010.767859] env[61768]: value = "vm-265539" [ 1010.767859] env[61768]: _type = "VirtualMachine" [ 1010.767859] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1010.768203] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c4bb3973-08a0-47d2-9663-fadc92043a2d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.776448] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease: (returnval){ [ 1010.776448] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523dc89c-d243-d34e-9a11-354b5b0de115" [ 1010.776448] env[61768]: _type = "HttpNfcLease" [ 1010.776448] env[61768]: } obtained for exporting VM: (result){ [ 1010.776448] env[61768]: value = "vm-265539" [ 1010.776448] env[61768]: _type = "VirtualMachine" [ 1010.776448] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1010.776728] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the lease: (returnval){ [ 1010.776728] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523dc89c-d243-d34e-9a11-354b5b0de115" [ 1010.776728] env[61768]: _type = "HttpNfcLease" [ 1010.776728] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1010.784131] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1010.784131] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523dc89c-d243-d34e-9a11-354b5b0de115" [ 1010.784131] env[61768]: _type = "HttpNfcLease" [ 1010.784131] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1010.786676] env[61768]: INFO nova.compute.manager [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Rescuing [ 1010.787050] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.787141] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.787285] env[61768]: DEBUG nova.network.neutron [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1010.884451] env[61768]: DEBUG oslo_vmware.api [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229406, 'name': ReconfigVM_Task, 'duration_secs': 0.31259} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.884870] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265537', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'name': 'volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'serial': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1010.997936] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.285936] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1011.285936] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523dc89c-d243-d34e-9a11-354b5b0de115" [ 1011.285936] env[61768]: _type = "HttpNfcLease" [ 1011.285936] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1011.286278] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1011.286278] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]523dc89c-d243-d34e-9a11-354b5b0de115" [ 1011.286278] env[61768]: _type = "HttpNfcLease" [ 1011.286278] env[61768]: }. {{(pid=61768) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1011.287020] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c139af73-9869-409a-a7c9-a51f0974f7a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.297521] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ef01d-d8c1-0339-d7c6-4af3a8b5cbcc/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1011.297801] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ef01d-d8c1-0339-d7c6-4af3a8b5cbcc/disk-0.vmdk for reading. {{(pid=61768) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1011.474214] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-e9d427ee-a252-4642-9515-146ee0d7846c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.610583] env[61768]: DEBUG nova.network.neutron [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updating instance_info_cache with network_info: [{"id": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "address": "fa:16:3e:b3:4f:18", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4f2467a-e3", "ovs_interfaceid": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.622692] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf26f5f7-750c-4ee9-9172-f5101b5c91f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.630846] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c16c94-260e-4912-a624-ca657a4f6097 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.662584] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c4acaf-2b9f-476f-b6f2-9c8f26c3c62e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.671739] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-876d2a14-cbbc-4f38-828e-2eae0741f685 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.685986] env[61768]: DEBUG nova.compute.provider_tree [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.933790] env[61768]: DEBUG nova.objects.instance [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid e1e90c98-5a65-473c-b508-b45fd93e31ad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.113433] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.190082] env[61768]: DEBUG nova.scheduler.client.report [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.439350] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d2016655-af11-41a2-b56a-e163670a438b tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.809s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.642273] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1012.643112] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d6cad4d7-9e6e-4530-8067-be1249026fcc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.652813] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1012.652813] env[61768]: value = "task-1229408" [ 1012.652813] env[61768]: _type = "Task" [ 1012.652813] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.662287] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229408, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.695883] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.230s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.696175] env[61768]: DEBUG nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1012.698910] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 1.701s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.166021] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229408, 'name': PowerOffVM_Task, 'duration_secs': 0.355281} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.166469] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1013.167651] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52ca2df-0813-410f-b05a-ddb31e997234 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.190833] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5449e2-6d32-44e2-9d61-bf1a7165ee32 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.202284] env[61768]: DEBUG nova.compute.utils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1013.205581] env[61768]: INFO nova.compute.claims [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1013.210199] env[61768]: DEBUG nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Not allocating networking since 'none' was specified. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 1013.223732] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1013.224551] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1bdca14d-de5e-4445-9783-aa5c191dd547 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.232506] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1013.232506] env[61768]: value = "task-1229409" [ 1013.232506] env[61768]: _type = "Task" [ 1013.232506] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.242092] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229409, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.248050] env[61768]: DEBUG oslo_concurrency.lockutils [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.248264] env[61768]: DEBUG oslo_concurrency.lockutils [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.711587] env[61768]: DEBUG nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1013.715960] env[61768]: INFO nova.compute.resource_tracker [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating resource usage from migration 3613c050-3dc4-4309-a637-fa3d3ec7535e [ 1013.746109] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1013.746376] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.746668] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.747041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.747041] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.747293] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fccbb1f3-8962-4f8c-b125-39427bcc7811 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.752256] env[61768]: INFO nova.compute.manager [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Detaching volume 31dc93bd-ffee-4257-90ed-c6adee08507a [ 1013.759521] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.759646] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1013.760667] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69ebca6e-5ff8-4db4-b6d0-0d2dcf5d6bb9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.772023] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1013.772023] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52afd12b-d99a-bb6c-34a0-07e2bba37878" [ 1013.772023] env[61768]: _type = "Task" [ 1013.772023] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.782660] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52afd12b-d99a-bb6c-34a0-07e2bba37878, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.800324] env[61768]: INFO nova.virt.block_device [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Attempting to driver detach volume 31dc93bd-ffee-4257-90ed-c6adee08507a from mountpoint /dev/sdb [ 1013.800540] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1013.800812] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265531', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'name': 'volume-31dc93bd-ffee-4257-90ed-c6adee08507a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'serial': '31dc93bd-ffee-4257-90ed-c6adee08507a'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1013.801828] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1f9939-334d-4b6f-aad8-e70dcfe21a6a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.833494] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ee65e1-0051-4786-ac8f-84076188a6b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.841402] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d220986-c7a1-4673-a5aa-aa67de6d5181 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.870155] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f20459c6-89f4-4387-941e-678f47a0d350 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.886893] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] The volume has not been displaced from its original location: [datastore1] volume-31dc93bd-ffee-4257-90ed-c6adee08507a/volume-31dc93bd-ffee-4257-90ed-c6adee08507a.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1013.892251] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfiguring VM instance instance-00000061 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1013.893575] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea8b931b-acf8-4d48-a0de-23d718cc22bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.907515] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662d62dd-c45e-4a66-831f-ae5080e02f27 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.915542] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7df15aed-f448-4c4e-bba4-d5443debe8ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.919918] env[61768]: DEBUG oslo_vmware.api [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1013.919918] env[61768]: value = "task-1229410" [ 1013.919918] env[61768]: _type = "Task" [ 1013.919918] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.950505] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e986048a-140d-43ca-9c5c-ee0b17d2e421 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.956531] env[61768]: DEBUG oslo_vmware.api [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229410, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.962520] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7bb3a0-26fd-4281-a889-72fe0e404e5f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.979646] env[61768]: DEBUG nova.compute.provider_tree [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.283979] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52afd12b-d99a-bb6c-34a0-07e2bba37878, 'name': SearchDatastore_Task, 'duration_secs': 0.014834} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.284844] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0819bd0c-5c80-4c26-8b90-e8e53f1c2028 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.291184] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1014.291184] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5273068c-f6a5-09f1-c78e-ce61897cbbcb" [ 1014.291184] env[61768]: _type = "Task" [ 1014.291184] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.300480] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5273068c-f6a5-09f1-c78e-ce61897cbbcb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.430978] env[61768]: DEBUG oslo_vmware.api [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229410, 'name': ReconfigVM_Task, 'duration_secs': 0.388889} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.431686] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfigured VM instance instance-00000061 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1014.436600] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e29ddf87-cb65-4105-ac96-895b7e5d57e4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.453639] env[61768]: DEBUG oslo_vmware.api [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1014.453639] env[61768]: value = "task-1229411" [ 1014.453639] env[61768]: _type = "Task" [ 1014.453639] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.464677] env[61768]: DEBUG oslo_vmware.api [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229411, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.483020] env[61768]: DEBUG nova.scheduler.client.report [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.729296] env[61768]: DEBUG nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1014.759766] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1014.760652] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1014.760652] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1014.760652] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1014.760823] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1014.761265] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1014.761379] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1014.761595] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1014.761851] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1014.762469] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1014.762550] env[61768]: DEBUG nova.virt.hardware [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1014.764312] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbdd0ed2-0961-473c-a7be-489a5db63705 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.775998] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d652d7-eaf4-4cb7-a581-debd7407db05 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.797428] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.807170] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Creating folder: Project (a934946fbd384208bd26b8b61e859454). Parent ref: group-v265360. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1014.812197] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-de5f84c5-fa4c-4423-87c7-aa1218b1b85d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.822742] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5273068c-f6a5-09f1-c78e-ce61897cbbcb, 'name': SearchDatastore_Task, 'duration_secs': 0.0153} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.823181] env[61768]: DEBUG oslo_concurrency.lockutils [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.823550] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. {{(pid=61768) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1014.825366] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-091d409a-5fa3-4b8b-a390-60c8310dcf12 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.828351] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Created folder: Project (a934946fbd384208bd26b8b61e859454) in parent group-v265360. [ 1014.828623] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Creating folder: Instances. Parent ref: group-v265540. {{(pid=61768) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1014.828948] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dce82c12-ad79-4e16-8d66-9c0c2505e005 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.837289] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1014.837289] env[61768]: value = "task-1229413" [ 1014.837289] env[61768]: _type = "Task" [ 1014.837289] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.842316] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Created folder: Instances in parent group-v265540. [ 1014.842737] env[61768]: DEBUG oslo.service.loopingcall [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.843273] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1014.843554] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e6c05efd-d324-45db-98c1-cdfd35535820 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.860297] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.865224] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.865224] env[61768]: value = "task-1229415" [ 1014.865224] env[61768]: _type = "Task" [ 1014.865224] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.873642] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229415, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.965921] env[61768]: DEBUG oslo_vmware.api [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229411, 'name': ReconfigVM_Task, 'duration_secs': 0.264756} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.966098] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265531', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'name': 'volume-31dc93bd-ffee-4257-90ed-c6adee08507a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': '31dc93bd-ffee-4257-90ed-c6adee08507a', 'serial': '31dc93bd-ffee-4257-90ed-c6adee08507a'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1014.987497] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.288s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.987671] env[61768]: INFO nova.compute.manager [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Migrating [ 1015.349085] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.376503] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229415, 'name': CreateVM_Task, 'duration_secs': 0.373038} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.376703] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1015.377167] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.377345] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.377704] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1015.377960] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12db2f7c-e089-437e-87e5-630cd21a8434 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.383259] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1015.383259] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f666e0-d0c9-b508-cd31-6c25d6e105da" [ 1015.383259] env[61768]: _type = "Task" [ 1015.383259] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.392018] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f666e0-d0c9-b508-cd31-6c25d6e105da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.505048] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.505274] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.505478] env[61768]: DEBUG nova.network.neutron [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1015.510458] env[61768]: DEBUG nova.objects.instance [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid e1e90c98-5a65-473c-b508-b45fd93e31ad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1015.848832] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.894815] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f666e0-d0c9-b508-cd31-6c25d6e105da, 'name': SearchDatastore_Task, 'duration_secs': 0.026155} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.895045] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.895336] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.895636] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.895803] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.895998] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.896344] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f3e8817-9bc0-43f0-92de-7a5a8bf23790 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.907599] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.907846] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1015.908665] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfe6cb18-152c-495f-958c-26ad98462e8a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.916532] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1015.916532] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520e3989-b629-03ac-6e37-34faefe8051c" [ 1015.916532] env[61768]: _type = "Task" [ 1015.916532] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.926894] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520e3989-b629-03ac-6e37-34faefe8051c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.328725] env[61768]: DEBUG nova.network.neutron [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance_info_cache with network_info: [{"id": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "address": "fa:16:3e:b9:94:f5", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8e6155-69", "ovs_interfaceid": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.352957] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229413, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.434798} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.353357] env[61768]: INFO nova.virt.vmwareapi.ds_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. [ 1016.354150] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f0ecc1f-dd2c-4bd5-9c3c-382a69ac728d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.382715] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.383073] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31bede2d-1b7b-4f6c-b237-d616b8f7eda1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.403267] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1016.403267] env[61768]: value = "task-1229416" [ 1016.403267] env[61768]: _type = "Task" [ 1016.403267] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.413102] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229416, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.427481] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520e3989-b629-03ac-6e37-34faefe8051c, 'name': SearchDatastore_Task, 'duration_secs': 0.023225} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.428491] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b908201d-9eaa-45bc-9a2b-798201b415b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.436025] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1016.436025] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f4a593-018a-0885-0986-998668199528" [ 1016.436025] env[61768]: _type = "Task" [ 1016.436025] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.446013] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f4a593-018a-0885-0986-998668199528, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.522094] env[61768]: DEBUG oslo_concurrency.lockutils [None req-83469e5c-6406-464d-a3aa-fd6fc5dff651 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.274s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.727327] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1016.727547] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.832428] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.915191] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229416, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.947662] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f4a593-018a-0885-0986-998668199528, 'name': SearchDatastore_Task, 'duration_secs': 0.017818} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.947949] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.948240] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1016.948530] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-87b4a0d8-760f-47e0-a49f-ef12f4761381 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.957278] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1016.957278] env[61768]: value = "task-1229417" [ 1016.957278] env[61768]: _type = "Task" [ 1016.957278] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.965929] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.231086] env[61768]: INFO nova.compute.manager [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Detaching volume da44c25e-cb5b-4c4c-94b3-b0af5fd7a004 [ 1017.269219] env[61768]: INFO nova.virt.block_device [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Attempting to driver detach volume da44c25e-cb5b-4c4c-94b3-b0af5fd7a004 from mountpoint /dev/sdc [ 1017.269662] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1017.269937] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265537', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'name': 'volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'serial': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1017.270908] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cf4ce15-1860-4829-988e-d0d9999ce84e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.295702] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a895ec5-d579-423f-8d8f-8ab901ba51ef {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.304036] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa442c9-8e30-4d7d-addd-eda953d23657 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.325706] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1379787-9bc1-45bb-9984-17bf9390a71f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.345763] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] The volume has not been displaced from its original location: [datastore2] volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004/volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1017.351248] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfiguring VM instance instance-00000061 to detach disk 2002 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1017.351883] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1413d67-dde5-4af7-b0c1-59fbfa3b1062 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.372256] env[61768]: DEBUG oslo_vmware.api [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1017.372256] env[61768]: value = "task-1229418" [ 1017.372256] env[61768]: _type = "Task" [ 1017.372256] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.382260] env[61768]: DEBUG oslo_vmware.api [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229418, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.415348] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229416, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.471381] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229417, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.884300] env[61768]: DEBUG oslo_vmware.api [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229418, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.915291] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229416, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.968871] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.785202} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.969218] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1017.969462] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.969723] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b5ef63aa-7e6b-402a-a2ce-47664637ec8c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.994935] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1017.994935] env[61768]: value = "task-1229419" [ 1017.994935] env[61768]: _type = "Task" [ 1017.994935] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.004246] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229419, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.359721] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae14f4b-e0bd-4bf2-ba12-197a41053e24 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.388296] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance 'a96d4721-83c1-474f-8470-a34262e34fad' progress to 0 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1018.406031] env[61768]: DEBUG oslo_vmware.api [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229418, 'name': ReconfigVM_Task, 'duration_secs': 0.578611} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.411133] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Reconfigured VM instance instance-00000061 to detach disk 2002 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1018.418385] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d63969c2-75fa-4356-9577-ee454d6ee09a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.438033] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229416, 'name': ReconfigVM_Task, 'duration_secs': 1.647013} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.439502] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Reconfigured VM instance instance-00000069 to attach disk [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.439948] env[61768]: DEBUG oslo_vmware.api [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1018.439948] env[61768]: value = "task-1229420" [ 1018.439948] env[61768]: _type = "Task" [ 1018.439948] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.440690] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829fd36e-5c3a-47fa-826f-2546b5172747 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.453043] env[61768]: DEBUG oslo_vmware.api [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229420, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.473814] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aae74096-2598-4229-9458-ad43aa48bb48 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.492488] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1018.492488] env[61768]: value = "task-1229421" [ 1018.492488] env[61768]: _type = "Task" [ 1018.492488] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.503929] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229419, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.101281} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.508125] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.508773] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229421, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.509564] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19944e20-b4c3-4794-a397-8658d687bd07 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.530984] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.531347] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9b2351f-e9f6-4277-a080-177efd8e62f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.552594] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1018.552594] env[61768]: value = "task-1229422" [ 1018.552594] env[61768]: _type = "Task" [ 1018.552594] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.561832] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229422, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.895689] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1018.896024] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-28c988cd-f2b7-475b-9c60-412e7d2ea60c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.904721] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1018.904721] env[61768]: value = "task-1229423" [ 1018.904721] env[61768]: _type = "Task" [ 1018.904721] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.915544] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229423, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.955620] env[61768]: DEBUG oslo_vmware.api [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229420, 'name': ReconfigVM_Task, 'duration_secs': 0.197538} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.956042] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265537', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'name': 'volume-da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'e1e90c98-5a65-473c-b508-b45fd93e31ad', 'attached_at': '', 'detached_at': '', 'volume_id': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004', 'serial': 'da44c25e-cb5b-4c4c-94b3-b0af5fd7a004'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1019.003839] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229421, 'name': ReconfigVM_Task, 'duration_secs': 0.199983} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.004203] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.004478] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b75eaffe-7c57-4d13-8fdf-d2503998139a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.012699] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1019.012699] env[61768]: value = "task-1229424" [ 1019.012699] env[61768]: _type = "Task" [ 1019.012699] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.023724] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229424, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.064069] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229422, 'name': ReconfigVM_Task, 'duration_secs': 0.295274} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.064382] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.065098] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfe3dc57-f7b3-445f-bdf4-508f8dc2e2fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.073205] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1019.073205] env[61768]: value = "task-1229425" [ 1019.073205] env[61768]: _type = "Task" [ 1019.073205] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.083144] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229425, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.414357] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229423, 'name': PowerOffVM_Task, 'duration_secs': 0.208589} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.414697] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1019.414839] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance 'a96d4721-83c1-474f-8470-a34262e34fad' progress to 17 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1019.505100] env[61768]: DEBUG nova.objects.instance [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'flavor' on Instance uuid e1e90c98-5a65-473c-b508-b45fd93e31ad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1019.527277] env[61768]: DEBUG oslo_vmware.api [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229424, 'name': PowerOnVM_Task, 'duration_secs': 0.410778} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.527631] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1019.530842] env[61768]: DEBUG nova.compute.manager [None req-8341591e-d4e6-40dd-ad99-838c9a78552a tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.532040] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ed611d8-e3d9-4b15-ad6a-28a6c25c6700 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.584163] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229425, 'name': Rename_Task, 'duration_secs': 0.139981} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.584424] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.584715] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-07eb6120-b4fd-4a62-9346-98003aea6696 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.592856] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1019.592856] env[61768]: value = "task-1229426" [ 1019.592856] env[61768]: _type = "Task" [ 1019.592856] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.601675] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229426, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.921504] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:42Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1019.921804] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1019.921994] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1019.922210] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1019.922373] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1019.922607] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1019.922911] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1019.923051] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1019.923239] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1019.923431] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1019.923627] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1019.928894] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b2dbb18-fa0f-4a45-a42b-6c7960627e60 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.947555] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1019.947555] env[61768]: value = "task-1229427" [ 1019.947555] env[61768]: _type = "Task" [ 1019.947555] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.957419] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229427, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.103574] env[61768]: DEBUG oslo_vmware.api [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229426, 'name': PowerOnVM_Task, 'duration_secs': 0.442883} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.103875] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1020.104114] env[61768]: INFO nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Took 5.37 seconds to spawn the instance on the hypervisor. [ 1020.104311] env[61768]: DEBUG nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.105143] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ca2e86-7e5d-4fcb-95f4-ee2ee3eefc74 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.458740] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229427, 'name': ReconfigVM_Task, 'duration_secs': 0.219828} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.459208] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance 'a96d4721-83c1-474f-8470-a34262e34fad' progress to 33 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1020.511323] env[61768]: DEBUG oslo_concurrency.lockutils [None req-9fd4abb9-d590-4024-9efd-e1576305bf2f tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.784s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.624462] env[61768]: INFO nova.compute.manager [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Took 10.18 seconds to build instance. [ 1020.965908] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.966216] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.966383] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.966575] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.966749] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.967179] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.967483] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.967670] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.967857] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.968047] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.968240] env[61768]: DEBUG nova.virt.hardware [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.973842] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Reconfiguring VM instance instance-0000006c to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1020.975090] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73f42fd5-9491-49cd-974f-3b87a79a1db4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.996870] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1020.996870] env[61768]: value = "task-1229428" [ 1020.996870] env[61768]: _type = "Task" [ 1020.996870] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.006776] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229428, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.126911] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a8a279da-455e-456f-9c0e-774c0b4e6098 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.693s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.143666] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.143903] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.144151] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "e1e90c98-5a65-473c-b508-b45fd93e31ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.144354] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.144540] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.146899] env[61768]: INFO nova.compute.manager [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Terminating instance [ 1021.148759] env[61768]: DEBUG nova.compute.manager [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.148968] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1021.149913] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a41a57-8634-433e-93f9-5cf45f3ecc1f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.158597] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1021.158855] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-124e9b15-0cd8-4c9b-b265-6541a7108fd3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.165844] env[61768]: DEBUG oslo_vmware.api [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1021.165844] env[61768]: value = "task-1229429" [ 1021.165844] env[61768]: _type = "Task" [ 1021.165844] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.175807] env[61768]: DEBUG oslo_vmware.api [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229429, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.198043] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ef01d-d8c1-0339-d7c6-4af3a8b5cbcc/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1021.199063] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a385436f-e50f-4964-999c-871ddc42865d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.205735] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ef01d-d8c1-0339-d7c6-4af3a8b5cbcc/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1021.205947] env[61768]: ERROR oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ef01d-d8c1-0339-d7c6-4af3a8b5cbcc/disk-0.vmdk due to incomplete transfer. [ 1021.206213] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-105effc0-469b-4be4-8d43-ed8644f99f22 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.223324] env[61768]: DEBUG oslo_vmware.rw_handles [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/528ef01d-d8c1-0339-d7c6-4af3a8b5cbcc/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1021.223574] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Uploaded image 9b2e869d-9e9c-4311-aae2-355101bc801a to the Glance image server {{(pid=61768) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1021.226182] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Destroying the VM {{(pid=61768) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1021.226502] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bf01e8d7-60ff-460c-ade5-ebb641b0d077 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.234629] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1021.234629] env[61768]: value = "task-1229430" [ 1021.234629] env[61768]: _type = "Task" [ 1021.234629] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.253036] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229430, 'name': Destroy_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.282028] env[61768]: INFO nova.compute.manager [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Rebuilding instance [ 1021.324431] env[61768]: DEBUG nova.compute.manager [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.325321] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f8bdad-646d-4f9f-a21e-ec1843d11f85 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.461496] env[61768]: INFO nova.compute.manager [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Unrescuing [ 1021.461828] env[61768]: DEBUG oslo_concurrency.lockutils [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.461990] env[61768]: DEBUG oslo_concurrency.lockutils [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.462180] env[61768]: DEBUG nova.network.neutron [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1021.507220] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229428, 'name': ReconfigVM_Task, 'duration_secs': 0.291454} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.507548] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Reconfigured VM instance instance-0000006c to detach disk 2000 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1021.508455] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64057794-b3b6-4987-b673-3ee65d205bd5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.534506] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] a96d4721-83c1-474f-8470-a34262e34fad/a96d4721-83c1-474f-8470-a34262e34fad.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1021.535242] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ae53b81-e328-49a3-a8e7-f0e0fdadb056 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.554460] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1021.554460] env[61768]: value = "task-1229431" [ 1021.554460] env[61768]: _type = "Task" [ 1021.554460] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.563685] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229431, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.676212] env[61768]: DEBUG oslo_vmware.api [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229429, 'name': PowerOffVM_Task, 'duration_secs': 0.236937} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.676523] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1021.676710] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1021.676979] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8e79d71-7cbe-4c87-916d-b2f03e277d9e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.745635] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229430, 'name': Destroy_Task, 'duration_secs': 0.402861} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.745902] env[61768]: INFO nova.virt.vmwareapi.vm_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Destroyed the VM [ 1021.746297] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Deleting Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1021.746640] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4e19e21b-fb23-4ade-b40e-86360c29e928 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.754616] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1021.754616] env[61768]: value = "task-1229433" [ 1021.754616] env[61768]: _type = "Task" [ 1021.754616] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.764465] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1021.764692] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1021.764874] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Deleting the datastore file [datastore2] e1e90c98-5a65-473c-b508-b45fd93e31ad {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1021.765173] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87d7b0f8-daef-41a8-a1b5-9f775d071263 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.770490] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229433, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.777320] env[61768]: DEBUG oslo_vmware.api [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for the task: (returnval){ [ 1021.777320] env[61768]: value = "task-1229434" [ 1021.777320] env[61768]: _type = "Task" [ 1021.777320] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.786647] env[61768]: DEBUG oslo_vmware.api [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229434, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.838383] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1021.838748] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18203384-fc6f-497d-a3bf-ffebcafb63e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.847313] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1021.847313] env[61768]: value = "task-1229435" [ 1021.847313] env[61768]: _type = "Task" [ 1021.847313] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.857018] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.066016] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229431, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.238834] env[61768]: DEBUG nova.network.neutron [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updating instance_info_cache with network_info: [{"id": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "address": "fa:16:3e:b3:4f:18", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4f2467a-e3", "ovs_interfaceid": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.265623] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229433, 'name': RemoveSnapshot_Task, 'duration_secs': 0.394854} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.265908] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Deleted Snapshot of the VM instance {{(pid=61768) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1022.266216] env[61768]: DEBUG nova.compute.manager [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.267024] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-950c87f4-38d7-4d38-a99c-f8da1b886ca1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.287725] env[61768]: DEBUG oslo_vmware.api [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Task: {'id': task-1229434, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.152125} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.287725] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.287725] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1022.288169] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1022.288169] env[61768]: INFO nova.compute.manager [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1022.291046] env[61768]: DEBUG oslo.service.loopingcall [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.291046] env[61768]: DEBUG nova.compute.manager [-] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.291046] env[61768]: DEBUG nova.network.neutron [-] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1022.357756] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229435, 'name': PowerOffVM_Task, 'duration_secs': 0.144847} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.358037] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1022.358268] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1022.359159] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-186852d3-1225-4022-9d15-07478105ae0c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.366890] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1022.367155] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-67418201-de3b-473e-ba37-5444888378d4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.394107] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1022.394351] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1022.394547] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Deleting the datastore file [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.394808] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-518c7348-d062-471b-baa8-32abfc83ef87 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.402537] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1022.402537] env[61768]: value = "task-1229437" [ 1022.402537] env[61768]: _type = "Task" [ 1022.402537] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.411690] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.565072] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229431, 'name': ReconfigVM_Task, 'duration_secs': 0.544151} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.565457] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Reconfigured VM instance instance-0000006c to attach disk [datastore2] a96d4721-83c1-474f-8470-a34262e34fad/a96d4721-83c1-474f-8470-a34262e34fad.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1022.565679] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance 'a96d4721-83c1-474f-8470-a34262e34fad' progress to 50 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1022.725215] env[61768]: DEBUG nova.compute.manager [req-ee43b5ed-5ef0-4789-ac33-675b5468ccc6 req-e34856a5-08e5-46c5-973c-68ef319ccc2f service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Received event network-vif-deleted-3ceeacd4-686d-4a24-bafc-328b8bd344ed {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1022.725668] env[61768]: INFO nova.compute.manager [req-ee43b5ed-5ef0-4789-ac33-675b5468ccc6 req-e34856a5-08e5-46c5-973c-68ef319ccc2f service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Neutron deleted interface 3ceeacd4-686d-4a24-bafc-328b8bd344ed; detaching it from the instance and deleting it from the info cache [ 1022.725668] env[61768]: DEBUG nova.network.neutron [req-ee43b5ed-5ef0-4789-ac33-675b5468ccc6 req-e34856a5-08e5-46c5-973c-68ef319ccc2f service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.741618] env[61768]: DEBUG oslo_concurrency.lockutils [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.742315] env[61768]: DEBUG nova.objects.instance [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lazy-loading 'flavor' on Instance uuid e8095d6d-18c5-4ca9-8ac8-475aeeffc01a {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.780211] env[61768]: INFO nova.compute.manager [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Shelve offloading [ 1022.781701] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1022.781960] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd9e2855-3d24-49a3-aac7-10b0f67a3194 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.790151] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1022.790151] env[61768]: value = "task-1229438" [ 1022.790151] env[61768]: _type = "Task" [ 1022.790151] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.799538] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229438, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.913631] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.452686} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.913907] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.914113] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1022.914297] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1023.072787] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ebcfd22-0aaa-4a17-ba42-f5614d71ea44 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.093131] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ffbf06b-6b3e-4837-be55-c563ca1d6eed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.111804] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance 'a96d4721-83c1-474f-8470-a34262e34fad' progress to 67 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1023.202816] env[61768]: DEBUG nova.network.neutron [-] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.228494] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ca4f91b-bb31-4a5f-ba6e-373715ba8655 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.239166] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c437e1-aecc-449a-9acb-7c6b8aa88484 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.257178] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2f9badb-dcc7-41d0-b553-de85bfcb3b34 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.278113] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.285243] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e92f747-8ef2-4288-86ff-9cc934d513bc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.287159] env[61768]: DEBUG nova.compute.manager [req-ee43b5ed-5ef0-4789-ac33-675b5468ccc6 req-e34856a5-08e5-46c5-973c-68ef319ccc2f service nova] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Detach interface failed, port_id=3ceeacd4-686d-4a24-bafc-328b8bd344ed, reason: Instance e1e90c98-5a65-473c-b508-b45fd93e31ad could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1023.296524] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1023.296524] env[61768]: value = "task-1229439" [ 1023.296524] env[61768]: _type = "Task" [ 1023.296524] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.305844] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1023.305844] env[61768]: DEBUG nova.compute.manager [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1023.306333] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec51bab6-f6d2-4799-8f63-81e8800ad775 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.312731] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.317256] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1023.317459] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.317689] env[61768]: DEBUG nova.network.neutron [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1023.654646] env[61768]: DEBUG nova.network.neutron [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Port 3c8e6155-6935-4025-a622-f7e4f08a0ec3 binding to destination host cpu-1 is already ACTIVE {{(pid=61768) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1023.705688] env[61768]: INFO nova.compute.manager [-] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Took 1.42 seconds to deallocate network for instance. [ 1023.811227] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229439, 'name': PowerOffVM_Task, 'duration_secs': 0.207374} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.811690] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1023.820282] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1023.822543] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1ed841c-6761-4e0d-896a-4947a23b7aed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.842336] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1023.842336] env[61768]: value = "task-1229440" [ 1023.842336] env[61768]: _type = "Task" [ 1023.842336] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.854058] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229440, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.947094] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.947418] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.947629] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.947874] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.948078] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.948288] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.948550] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.948803] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.949109] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.949372] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.949642] env[61768]: DEBUG nova.virt.hardware [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.950856] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea1bd59-15a1-4cc0-af06-c0a3207c633a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.960906] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246a8740-05da-4192-b12f-ce47e3a63107 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.975785] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Instance VIF info [] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.981226] env[61768]: DEBUG oslo.service.loopingcall [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.981462] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1023.981699] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ec770c4-7a65-4c43-8612-98f1da2ab4eb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.001900] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1024.001900] env[61768]: value = "task-1229441" [ 1024.001900] env[61768]: _type = "Task" [ 1024.001900] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.010606] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229441, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.082134] env[61768]: DEBUG nova.network.neutron [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.211622] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.211967] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.212240] env[61768]: DEBUG nova.objects.instance [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lazy-loading 'resources' on Instance uuid e1e90c98-5a65-473c-b508-b45fd93e31ad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1024.353816] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229440, 'name': ReconfigVM_Task, 'duration_secs': 0.229169} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.354178] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1024.354382] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1024.354644] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dd18bbdc-a0b1-413e-a6ea-c69ea70905f4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.363274] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1024.363274] env[61768]: value = "task-1229442" [ 1024.363274] env[61768]: _type = "Task" [ 1024.363274] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.372577] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229442, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.512752] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229441, 'name': CreateVM_Task, 'duration_secs': 0.271398} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.512954] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.513444] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.513624] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.514019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.514304] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da27ed9c-c6db-4ee7-b3aa-68b0811dc3f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.520235] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1024.520235] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f82586-42ca-0933-076a-c71991b0cd9c" [ 1024.520235] env[61768]: _type = "Task" [ 1024.520235] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.529119] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f82586-42ca-0933-076a-c71991b0cd9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.585113] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.678578] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "a96d4721-83c1-474f-8470-a34262e34fad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.678912] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.679162] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.829216] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4a449be-61c8-4846-a7ee-f11dbfd3ca61 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.836387] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3494e478-8de8-4f05-8656-9eadb60cedd9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.872949] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0779321-6189-4602-a9c5-30ff338a49f7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.876347] env[61768]: DEBUG nova.compute.manager [req-0348ad75-899d-4a4f-9cbb-5e8bd3a89ea2 req-ed57bd6e-be0a-4583-ab57-85f9d7e3f90b service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-vif-unplugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1024.876562] env[61768]: DEBUG oslo_concurrency.lockutils [req-0348ad75-899d-4a4f-9cbb-5e8bd3a89ea2 req-ed57bd6e-be0a-4583-ab57-85f9d7e3f90b service nova] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1024.876774] env[61768]: DEBUG oslo_concurrency.lockutils [req-0348ad75-899d-4a4f-9cbb-5e8bd3a89ea2 req-ed57bd6e-be0a-4583-ab57-85f9d7e3f90b service nova] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.876945] env[61768]: DEBUG oslo_concurrency.lockutils [req-0348ad75-899d-4a4f-9cbb-5e8bd3a89ea2 req-ed57bd6e-be0a-4583-ab57-85f9d7e3f90b service nova] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.877136] env[61768]: DEBUG nova.compute.manager [req-0348ad75-899d-4a4f-9cbb-5e8bd3a89ea2 req-ed57bd6e-be0a-4583-ab57-85f9d7e3f90b service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] No waiting events found dispatching network-vif-unplugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1024.877314] env[61768]: WARNING nova.compute.manager [req-0348ad75-899d-4a4f-9cbb-5e8bd3a89ea2 req-ed57bd6e-be0a-4583-ab57-85f9d7e3f90b service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received unexpected event network-vif-unplugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 for instance with vm_state shelved and task_state shelving_offloading. [ 1024.886593] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7955865-8c5c-4cd8-9096-89408cdb46cd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.891013] env[61768]: DEBUG oslo_vmware.api [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229442, 'name': PowerOnVM_Task, 'duration_secs': 0.388927} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.891587] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1024.891833] env[61768]: DEBUG nova.compute.manager [None req-32205638-518a-48b0-9ab3-c1e179a810bb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.893050] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6b9691-91e4-4d78-8f92-3c097a3c9436 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.904168] env[61768]: DEBUG nova.compute.provider_tree [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1024.945418] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1024.947251] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d30a96-4966-4124-a57e-06c2358d481f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.956897] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1024.957204] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c0b2ed50-33f8-41fd-ba12-302e49cbdcb9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.032625] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f82586-42ca-0933-076a-c71991b0cd9c, 'name': SearchDatastore_Task, 'duration_secs': 0.009853} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.033084] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.033483] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1025.033774] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.033951] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.034156] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1025.034583] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-680b3f65-faa2-4e11-9e5b-d9263c6f1bd8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.039798] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1025.040010] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1025.040214] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleting the datastore file [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.040459] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-63553230-ace8-465b-8b4e-0f617ae6b4c7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.046172] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1025.046440] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1025.048160] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06d11825-10d9-4395-ab47-7939fb7531c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.050607] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1025.050607] env[61768]: value = "task-1229444" [ 1025.050607] env[61768]: _type = "Task" [ 1025.050607] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.056080] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1025.056080] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52810588-aa62-13b0-81a9-f74bcbebcd3f" [ 1025.056080] env[61768]: _type = "Task" [ 1025.056080] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.062926] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229444, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.068914] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52810588-aa62-13b0-81a9-f74bcbebcd3f, 'name': SearchDatastore_Task, 'duration_secs': 0.00864} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.069745] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-329343d8-d08b-424e-b9b8-82919394c7b7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.075692] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1025.075692] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52076112-3f77-d079-2d77-e890eff20b13" [ 1025.075692] env[61768]: _type = "Task" [ 1025.075692] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.085759] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52076112-3f77-d079-2d77-e890eff20b13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.406608] env[61768]: DEBUG nova.scheduler.client.report [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.560410] env[61768]: DEBUG oslo_vmware.api [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229444, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141084} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.560527] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.560724] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1025.560908] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1025.580665] env[61768]: INFO nova.scheduler.client.report [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted allocations for instance 81641ce1-cc88-4df1-a54c-48f9bd4bb73b [ 1025.589656] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52076112-3f77-d079-2d77-e890eff20b13, 'name': SearchDatastore_Task, 'duration_secs': 0.009435} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.589936] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1025.590219] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1025.590495] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea8e20a0-c680-44c3-b429-4429e6757888 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.597942] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1025.597942] env[61768]: value = "task-1229445" [ 1025.597942] env[61768]: _type = "Task" [ 1025.597942] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.606393] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229445, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.720337] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.720602] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.720816] env[61768]: DEBUG nova.network.neutron [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1025.788328] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "86f41ef0-4c92-4626-bb83-1148144c597c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.788584] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.912640] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.933641] env[61768]: INFO nova.scheduler.client.report [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Deleted allocations for instance e1e90c98-5a65-473c-b508-b45fd93e31ad [ 1026.085638] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.085925] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.086167] env[61768]: DEBUG nova.objects.instance [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'resources' on Instance uuid 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.108510] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229445, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474803} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.108768] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1026.108993] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1026.109258] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-75611ebc-6786-44c7-bd29-10e524d55ed1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.116371] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1026.116371] env[61768]: value = "task-1229446" [ 1026.116371] env[61768]: _type = "Task" [ 1026.116371] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.124476] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229446, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.292182] env[61768]: DEBUG nova.compute.utils [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.432699] env[61768]: DEBUG nova.network.neutron [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance_info_cache with network_info: [{"id": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "address": "fa:16:3e:b9:94:f5", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8e6155-69", "ovs_interfaceid": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.440646] env[61768]: DEBUG oslo_concurrency.lockutils [None req-45a5092e-5cc7-457a-b2b5-3c561bf0e188 tempest-AttachVolumeTestJSON-1106731130 tempest-AttachVolumeTestJSON-1106731130-project-member] Lock "e1e90c98-5a65-473c-b508-b45fd93e31ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.297s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.589590] env[61768]: DEBUG nova.objects.instance [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'numa_topology' on Instance uuid 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1026.630097] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229446, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.093898} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.630097] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.630097] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0e3983-9632-4f03-a48b-23e69c84332c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.650246] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Reconfiguring VM instance instance-0000006d to attach disk [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.650924] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb19f9cb-3f59-43a2-8d53-041a4c1ab0a6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.671674] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1026.671674] env[61768]: value = "task-1229448" [ 1026.671674] env[61768]: _type = "Task" [ 1026.671674] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.680352] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229448, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.795722] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.936475] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.091281] env[61768]: DEBUG nova.objects.base [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Object Instance<81641ce1-cc88-4df1-a54c-48f9bd4bb73b> lazy-loaded attributes: resources,numa_topology {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1027.186663] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229448, 'name': ReconfigVM_Task, 'duration_secs': 0.271549} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.187750] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Reconfigured VM instance instance-0000006d to attach disk [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26/0e598e1b-754d-4a2f-92e6-2e43463d9e26.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.187750] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57692dff-91bd-45f9-af54-714fd99a3a0e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.190812] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-805b2221-f5be-4717-b926-ddb25a5c74a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.200566] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf2dcd0-d7ac-45e7-9f5c-7652bfeeafa3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.204369] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1027.204369] env[61768]: value = "task-1229449" [ 1027.204369] env[61768]: _type = "Task" [ 1027.204369] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.234818] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf257a5e-4fff-4a3d-92a2-ae71156c08ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.238696] env[61768]: DEBUG nova.compute.manager [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.238805] env[61768]: DEBUG nova.compute.manager [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing instance network info cache due to event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1027.239045] env[61768]: DEBUG oslo_concurrency.lockutils [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.239195] env[61768]: DEBUG oslo_concurrency.lockutils [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.239356] env[61768]: DEBUG nova.network.neutron [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing network info cache for port 8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1027.244654] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229449, 'name': Rename_Task} progress is 14%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.251326] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3ebd63-3ade-4b14-af45-dac976870bdd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.268843] env[61768]: DEBUG nova.compute.provider_tree [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.457692] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1fac0a5-6d3c-4d3e-be62-aff7bd6d8b24 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.477560] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76abcf6d-8be4-441d-94ff-8e8f74912163 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.485262] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance 'a96d4721-83c1-474f-8470-a34262e34fad' progress to 83 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1027.714692] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229449, 'name': Rename_Task, 'duration_secs': 0.159128} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.714979] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.715264] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b79279d1-60c9-4f34-96a6-05e22b8f428d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.723506] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1027.723506] env[61768]: value = "task-1229450" [ 1027.723506] env[61768]: _type = "Task" [ 1027.723506] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.734964] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229450, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.774843] env[61768]: DEBUG nova.scheduler.client.report [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1027.858162] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "86f41ef0-4c92-4626-bb83-1148144c597c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.858513] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.858674] env[61768]: INFO nova.compute.manager [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Attaching volume fc081bbb-c037-479f-8460-545c7cfb7e0d to /dev/sdb [ 1027.892291] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320e5e20-8df9-4f03-80bc-4288dc3a8e49 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.902136] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef120e32-82e8-4315-aa21-f004d42ce16a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.916947] env[61768]: DEBUG nova.virt.block_device [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updating existing volume attachment record: dfffdf30-2f6a-4906-91da-e22b98a31328 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1027.991101] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.991423] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d29019a1-5e05-4550-912c-0fe748da402b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.999680] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1027.999680] env[61768]: value = "task-1229451" [ 1027.999680] env[61768]: _type = "Task" [ 1027.999680] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.008751] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229451, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.098829] env[61768]: DEBUG nova.network.neutron [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updated VIF entry in instance network info cache for port 8e676cc0-ee67-4e76-afe1-2dff2e254617. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1028.099436] env[61768]: DEBUG nova.network.neutron [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.234391] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229450, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.280316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.194s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.509723] env[61768]: DEBUG oslo_vmware.api [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229451, 'name': PowerOnVM_Task, 'duration_secs': 0.487294} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.510030] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.510235] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-a9c5fa68-cb88-42f2-8ca3-6a14169ab87a tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance 'a96d4721-83c1-474f-8470-a34262e34fad' progress to 100 {{(pid=61768) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1028.602251] env[61768]: DEBUG oslo_concurrency.lockutils [req-43c1674e-c2cc-4d3c-852c-4db238c6e0a4 req-a5718c65-8c57-4ae9-a418-515a8382ef16 service nova] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.734498] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229450, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.788381] env[61768]: DEBUG oslo_concurrency.lockutils [None req-a6d0e4f2-a746-4299-91ad-a6e119e6cd20 tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.706s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.023543] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.023830] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.182872] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.183180] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.183374] env[61768]: INFO nova.compute.manager [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Unshelving [ 1029.235091] env[61768]: DEBUG oslo_vmware.api [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229450, 'name': PowerOnVM_Task, 'duration_secs': 1.031118} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.235371] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1029.235579] env[61768]: DEBUG nova.compute.manager [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1029.236444] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3fa33c-fd1d-4059-a3af-82824933b81d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.525840] env[61768]: DEBUG nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1029.751329] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.751817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1029.751864] env[61768]: DEBUG nova.objects.instance [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61768) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1030.057482] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.205167] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.762068] env[61768]: DEBUG oslo_concurrency.lockutils [None req-676c71ce-7e0b-4251-a2fe-eaded390cd3d tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.763352] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.763593] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.763796] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.763992] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.764186] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.765725] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.708s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.767202] env[61768]: INFO nova.compute.claims [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1030.769746] env[61768]: INFO nova.compute.manager [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Terminating instance [ 1030.773817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "refresh_cache-0e598e1b-754d-4a2f-92e6-2e43463d9e26" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.773981] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquired lock "refresh_cache-0e598e1b-754d-4a2f-92e6-2e43463d9e26" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.774239] env[61768]: DEBUG nova.network.neutron [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1030.903099] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "a96d4721-83c1-474f-8470-a34262e34fad" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.903388] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.903586] env[61768]: DEBUG nova.compute.manager [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Going to confirm migration 3 {{(pid=61768) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1031.162258] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.162745] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1031.162745] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1031.291918] env[61768]: DEBUG nova.network.neutron [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1031.339120] env[61768]: DEBUG nova.network.neutron [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.471718] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.471944] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquired lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.472143] env[61768]: DEBUG nova.network.neutron [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1031.472339] env[61768]: DEBUG nova.objects.instance [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'info_cache' on Instance uuid a96d4721-83c1-474f-8470-a34262e34fad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.841762] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Releasing lock "refresh_cache-0e598e1b-754d-4a2f-92e6-2e43463d9e26" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.842266] env[61768]: DEBUG nova.compute.manager [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.842513] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1031.843431] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e261b5c5-5a38-4022-869f-b14f0f920cd6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.851243] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1031.851510] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-55b4d2d1-5789-4e39-922f-8982f5659196 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.859417] env[61768]: DEBUG oslo_vmware.api [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1031.859417] env[61768]: value = "task-1229455" [ 1031.859417] env[61768]: _type = "Task" [ 1031.859417] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.866983] env[61768]: DEBUG oslo_vmware.api [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229455, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.878660] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40eeff07-81f4-4630-86de-92709e9182b5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.885290] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f121b0a6-3264-446c-9e5a-486e9864afd5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.914847] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e35572d0-41ce-4f04-b572-4a0fec4e6999 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.921741] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd0e093-d590-4fd8-9ad7-ac7287e8d035 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.935190] env[61768]: DEBUG nova.compute.provider_tree [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.369378] env[61768]: DEBUG oslo_vmware.api [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229455, 'name': PowerOffVM_Task, 'duration_secs': 0.120133} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.369709] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1032.369830] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1032.370096] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c9a2149-0929-4758-a6c2-bbd9b335a912 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.395997] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1032.396286] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1032.396484] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Deleting the datastore file [datastore2] 0e598e1b-754d-4a2f-92e6-2e43463d9e26 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.396762] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-82aaebf3-e8f8-4146-9894-20b1b6bb4ee5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.403023] env[61768]: DEBUG oslo_vmware.api [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for the task: (returnval){ [ 1032.403023] env[61768]: value = "task-1229457" [ 1032.403023] env[61768]: _type = "Task" [ 1032.403023] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.410791] env[61768]: DEBUG oslo_vmware.api [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229457, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.438976] env[61768]: DEBUG nova.scheduler.client.report [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.468555] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1032.468906] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265544', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'name': 'volume-fc081bbb-c037-479f-8460-545c7cfb7e0d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86f41ef0-4c92-4626-bb83-1148144c597c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'serial': 'fc081bbb-c037-479f-8460-545c7cfb7e0d'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1032.470059] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dd131aa-40a5-4a2a-a73f-d84eb7d1b8bd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.488696] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d48dea-1317-4b31-8e64-ba7cfe71cc81 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.514491] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] volume-fc081bbb-c037-479f-8460-545c7cfb7e0d/volume-fc081bbb-c037-479f-8460-545c7cfb7e0d.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1032.514823] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9685f2ab-14b0-4033-8813-83c9a635079c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.534807] env[61768]: DEBUG oslo_vmware.api [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1032.534807] env[61768]: value = "task-1229458" [ 1032.534807] env[61768]: _type = "Task" [ 1032.534807] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.543029] env[61768]: DEBUG oslo_vmware.api [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229458, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.744338] env[61768]: DEBUG nova.network.neutron [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance_info_cache with network_info: [{"id": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "address": "fa:16:3e:b9:94:f5", "network": {"id": "876affa4-f1aa-4b58-8898-6560e5110f7f", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1995398005-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dfce0cf4a5d14042a929dac65aeb2ae3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "98e21102-8954-4f6f-b1e6-5d764a53aa22", "external-id": "nsx-vlan-transportzone-838", "segmentation_id": 838, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c8e6155-69", "ovs_interfaceid": "3c8e6155-6935-4025-a622-f7e4f08a0ec3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.912623] env[61768]: DEBUG oslo_vmware.api [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Task: {'id': task-1229457, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096733} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.912901] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1032.913106] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1032.913326] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1032.913524] env[61768]: INFO nova.compute.manager [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Took 1.07 seconds to destroy the instance on the hypervisor. [ 1032.913773] env[61768]: DEBUG oslo.service.loopingcall [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.913975] env[61768]: DEBUG nova.compute.manager [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1032.914092] env[61768]: DEBUG nova.network.neutron [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1032.929443] env[61768]: DEBUG nova.network.neutron [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1032.943536] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.944361] env[61768]: DEBUG nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1032.950020] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.745s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.950020] env[61768]: DEBUG nova.objects.instance [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'pci_requests' on Instance uuid 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.044970] env[61768]: DEBUG oslo_vmware.api [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229458, 'name': ReconfigVM_Task, 'duration_secs': 0.355724} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.045294] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Reconfigured VM instance instance-00000067 to attach disk [datastore2] volume-fc081bbb-c037-479f-8460-545c7cfb7e0d/volume-fc081bbb-c037-479f-8460-545c7cfb7e0d.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1033.050150] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ed3daca-582f-4733-97c6-a17f6a0ae725 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.064810] env[61768]: DEBUG oslo_vmware.api [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1033.064810] env[61768]: value = "task-1229460" [ 1033.064810] env[61768]: _type = "Task" [ 1033.064810] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.072352] env[61768]: DEBUG oslo_vmware.api [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229460, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.246652] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Releasing lock "refresh_cache-a96d4721-83c1-474f-8470-a34262e34fad" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.246883] env[61768]: DEBUG nova.objects.instance [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'migration_context' on Instance uuid a96d4721-83c1-474f-8470-a34262e34fad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.432182] env[61768]: DEBUG nova.network.neutron [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.456011] env[61768]: DEBUG nova.objects.instance [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'numa_topology' on Instance uuid 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.457450] env[61768]: DEBUG nova.compute.utils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1033.458734] env[61768]: DEBUG nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1033.458903] env[61768]: DEBUG nova.network.neutron [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1033.518413] env[61768]: DEBUG nova.policy [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26e3d09d20274851950ae324bcf85a14', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e4c2ff63d9d94a41b0dbc7d597e762a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 1033.575076] env[61768]: DEBUG oslo_vmware.api [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229460, 'name': ReconfigVM_Task, 'duration_secs': 0.133784} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.575460] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265544', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'name': 'volume-fc081bbb-c037-479f-8460-545c7cfb7e0d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86f41ef0-4c92-4626-bb83-1148144c597c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'serial': 'fc081bbb-c037-479f-8460-545c7cfb7e0d'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1033.749331] env[61768]: DEBUG nova.objects.base [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61768) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1033.750327] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b13a33a0-df38-49f3-a153-3a04f3030c4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.769535] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf2259b3-c5b8-4d1f-a0cc-cef50ee87e7f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.775580] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1033.775580] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5244f593-5bfc-e75e-b87c-e8379f62ba3c" [ 1033.775580] env[61768]: _type = "Task" [ 1033.775580] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.784383] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5244f593-5bfc-e75e-b87c-e8379f62ba3c, 'name': SearchDatastore_Task, 'duration_secs': 0.006681} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.784711] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.805641] env[61768]: DEBUG nova.network.neutron [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Successfully created port: 3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1033.934945] env[61768]: INFO nova.compute.manager [-] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Took 1.02 seconds to deallocate network for instance. [ 1033.960349] env[61768]: INFO nova.compute.claims [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.968642] env[61768]: DEBUG nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1034.443775] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.612972] env[61768]: DEBUG nova.objects.instance [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'flavor' on Instance uuid 86f41ef0-4c92-4626-bb83-1148144c597c {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.682821] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Didn't find any instances for network info cache update. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1034.683040] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.683215] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.683366] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.683521] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.683818] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.683818] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.683999] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1034.684187] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1034.980501] env[61768]: DEBUG nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1035.007045] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1035.007369] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1035.007478] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1035.007664] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1035.007818] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1035.007974] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1035.008276] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1035.008469] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1035.008648] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1035.008818] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1035.008996] env[61768]: DEBUG nova.virt.hardware [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1035.009886] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-003c1a4a-a64f-4a4e-8cfe-dca7081e6ff5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.020045] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b19882e3-6874-4b36-b4bd-8d14e2e07cbd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.093579] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8900c863-da90-4559-80fb-c766ec9fa194 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.101431] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-817765e1-30aa-4c21-85bc-0fce0b1da624 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.133569] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c8342a-e836-421e-b81f-4d49979ddb17 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.136543] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2ba23791-524d-4f80-858e-e95fc4b14e80 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.278s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.142618] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d15f8f-ff3d-4eac-87c4-3ece22393086 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.156516] env[61768]: DEBUG nova.compute.provider_tree [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.187566] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.337041] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "86f41ef0-4c92-4626-bb83-1148144c597c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.337297] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.372850] env[61768]: DEBUG nova.compute.manager [req-b3806f1e-265e-4113-9f0a-d29330bebaa6 req-51d1687b-2cb0-480f-874e-c9f1aace0dea service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Received event network-vif-plugged-3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1035.373104] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3806f1e-265e-4113-9f0a-d29330bebaa6 req-51d1687b-2cb0-480f-874e-c9f1aace0dea service nova] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.373322] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3806f1e-265e-4113-9f0a-d29330bebaa6 req-51d1687b-2cb0-480f-874e-c9f1aace0dea service nova] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.373525] env[61768]: DEBUG oslo_concurrency.lockutils [req-b3806f1e-265e-4113-9f0a-d29330bebaa6 req-51d1687b-2cb0-480f-874e-c9f1aace0dea service nova] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.373655] env[61768]: DEBUG nova.compute.manager [req-b3806f1e-265e-4113-9f0a-d29330bebaa6 req-51d1687b-2cb0-480f-874e-c9f1aace0dea service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] No waiting events found dispatching network-vif-plugged-3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1035.373824] env[61768]: WARNING nova.compute.manager [req-b3806f1e-265e-4113-9f0a-d29330bebaa6 req-51d1687b-2cb0-480f-874e-c9f1aace0dea service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Received unexpected event network-vif-plugged-3770838d-f101-43db-99f9-fc4b33c5bda7 for instance with vm_state building and task_state spawning. [ 1035.480657] env[61768]: DEBUG nova.network.neutron [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Successfully updated port: 3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1035.659859] env[61768]: DEBUG nova.scheduler.client.report [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.842198] env[61768]: INFO nova.compute.manager [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Detaching volume fc081bbb-c037-479f-8460-545c7cfb7e0d [ 1035.884356] env[61768]: INFO nova.virt.block_device [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Attempting to driver detach volume fc081bbb-c037-479f-8460-545c7cfb7e0d from mountpoint /dev/sdb [ 1035.884771] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1035.885115] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265544', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'name': 'volume-fc081bbb-c037-479f-8460-545c7cfb7e0d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86f41ef0-4c92-4626-bb83-1148144c597c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'serial': 'fc081bbb-c037-479f-8460-545c7cfb7e0d'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1035.886448] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2d1ebf-7c71-4756-ae34-aa2ceb54a43f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.914376] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eb6d829-1a44-4b6e-b81e-2a9345c609a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.921824] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-555c499d-5c5a-4acb-b0f1-db5679d858dd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.942542] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3a016c5-2c99-4cc4-b14a-7de09193e6c0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.958276] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] The volume has not been displaced from its original location: [datastore2] volume-fc081bbb-c037-479f-8460-545c7cfb7e0d/volume-fc081bbb-c037-479f-8460-545c7cfb7e0d.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1035.964057] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Reconfiguring VM instance instance-00000067 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1035.964380] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c187408d-b2f0-4f49-bcfa-f40c08b78729 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.983306] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.983493] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.983668] env[61768]: DEBUG nova.network.neutron [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1035.988418] env[61768]: DEBUG oslo_vmware.api [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1035.988418] env[61768]: value = "task-1229461" [ 1035.988418] env[61768]: _type = "Task" [ 1035.988418] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.997285] env[61768]: DEBUG oslo_vmware.api [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229461, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.165554] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.216s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.167759] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 2.383s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.239177] env[61768]: INFO nova.network.neutron [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating port 8e676cc0-ee67-4e76-afe1-2dff2e254617 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1036.498341] env[61768]: DEBUG oslo_vmware.api [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229461, 'name': ReconfigVM_Task, 'duration_secs': 0.406329} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.498783] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Reconfigured VM instance instance-00000067 to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1036.504157] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24a8365f-5165-47a3-be3e-97ebc700234b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.520477] env[61768]: DEBUG oslo_vmware.api [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1036.520477] env[61768]: value = "task-1229462" [ 1036.520477] env[61768]: _type = "Task" [ 1036.520477] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.528553] env[61768]: DEBUG oslo_vmware.api [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229462, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.557844] env[61768]: DEBUG nova.network.neutron [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1036.805692] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f5203be-bb8c-4817-8dac-fcbe2bb4b4cb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.813771] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c07939-612c-4ae1-85df-10b3c20057ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.843433] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90cbcbbd-2c96-4f32-98c0-f652e0f35bed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.850550] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6b2b72-c4fe-452c-b1ad-1ce9a1da8b1d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.863924] env[61768]: DEBUG nova.compute.provider_tree [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.866846] env[61768]: DEBUG nova.network.neutron [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1037.030073] env[61768]: DEBUG oslo_vmware.api [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229462, 'name': ReconfigVM_Task, 'duration_secs': 0.149184} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.030874] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265544', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'name': 'volume-fc081bbb-c037-479f-8460-545c7cfb7e0d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '86f41ef0-4c92-4626-bb83-1148144c597c', 'attached_at': '', 'detached_at': '', 'volume_id': 'fc081bbb-c037-479f-8460-545c7cfb7e0d', 'serial': 'fc081bbb-c037-479f-8460-545c7cfb7e0d'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1037.367362] env[61768]: DEBUG nova.scheduler.client.report [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.371771] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.371771] env[61768]: DEBUG nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Instance network_info: |[{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1037.373066] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:81:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4162774e-ec80-4d85-aeb4-fae77f197393', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3770838d-f101-43db-99f9-fc4b33c5bda7', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1037.380083] env[61768]: DEBUG oslo.service.loopingcall [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1037.380280] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1037.380531] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06854711-6bfd-467c-a863-b9231fdf4d68 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.404718] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1037.404718] env[61768]: value = "task-1229463" [ 1037.404718] env[61768]: _type = "Task" [ 1037.404718] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.412378] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229463, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.455021] env[61768]: DEBUG nova.compute.manager [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Received event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.455251] env[61768]: DEBUG nova.compute.manager [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing instance network info cache due to event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.455485] env[61768]: DEBUG oslo_concurrency.lockutils [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] Acquiring lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.455641] env[61768]: DEBUG oslo_concurrency.lockutils [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] Acquired lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.455801] env[61768]: DEBUG nova.network.neutron [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1037.583247] env[61768]: DEBUG nova.objects.instance [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'flavor' on Instance uuid 86f41ef0-4c92-4626-bb83-1148144c597c {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.915457] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229463, 'name': CreateVM_Task, 'duration_secs': 0.307816} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.915643] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1037.916382] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.916567] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.916960] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1037.917203] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a080c47d-a09d-43e4-8384-494333d1309a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.922050] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1037.922050] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522f0947-e13b-ced2-7ee2-1772e3988e39" [ 1037.922050] env[61768]: _type = "Task" [ 1037.922050] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.929734] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522f0947-e13b-ced2-7ee2-1772e3988e39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.288136] env[61768]: DEBUG nova.network.neutron [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updated VIF entry in instance network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1038.288541] env[61768]: DEBUG nova.network.neutron [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.301518] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.302116] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.302116] env[61768]: DEBUG nova.network.neutron [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1038.379373] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.212s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.383308] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.939s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.383308] env[61768]: DEBUG nova.objects.instance [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lazy-loading 'resources' on Instance uuid 0e598e1b-754d-4a2f-92e6-2e43463d9e26 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1038.433649] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]522f0947-e13b-ced2-7ee2-1772e3988e39, 'name': SearchDatastore_Task, 'duration_secs': 0.009039} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.434737] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.434994] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1038.435290] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.435465] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.435659] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1038.436161] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-07441f7b-dcc3-4216-b024-caca2c359e7c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.444069] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1038.444255] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1038.444942] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9eb1bb39-220a-4673-9c42-d52ae1f34d97 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.449948] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1038.449948] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52becd65-8cc5-d0a8-5928-741ca94e8da9" [ 1038.449948] env[61768]: _type = "Task" [ 1038.449948] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.457534] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52becd65-8cc5-d0a8-5928-741ca94e8da9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.592592] env[61768]: DEBUG oslo_concurrency.lockutils [None req-5412be13-2062-4603-9baa-6ff90c9d6578 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.255s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.791668] env[61768]: DEBUG oslo_concurrency.lockutils [req-66601c7c-04e4-4c3b-a049-5a7563096edd req-041c55ec-930e-44d8-a237-39d36e587e7c service nova] Releasing lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.938956] env[61768]: INFO nova.scheduler.client.report [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted allocation for migration 3613c050-3dc4-4309-a637-fa3d3ec7535e [ 1038.965019] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52becd65-8cc5-d0a8-5928-741ca94e8da9, 'name': SearchDatastore_Task, 'duration_secs': 0.00826} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.965019] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4e1d6b3-bfde-4023-85e7-cdf61e0ca6ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.971279] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1038.971279] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520d8276-0a5d-edf9-67bd-3a6ebdd666a6" [ 1038.971279] env[61768]: _type = "Task" [ 1038.971279] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.990013] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520d8276-0a5d-edf9-67bd-3a6ebdd666a6, 'name': SearchDatastore_Task, 'duration_secs': 0.015644} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.992609] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1038.992985] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/c4a4d1d5-d91f-4513-9dff-c0a5398ead5e.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1038.993457] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f04d010d-95be-49ff-9ba7-efe401f2cbb7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.000605] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1039.000605] env[61768]: value = "task-1229464" [ 1039.000605] env[61768]: _type = "Task" [ 1039.000605] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.006082] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d0d8bc-d134-41d2-8b05-ce2af9c726ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.011757] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229464, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.016214] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fc503b-bf77-472a-8429-d51512e88864 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.045750] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6a0d43b-eb75-432f-bb2a-f75307324de9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.053088] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b9751f-06f3-4a4f-9ca0-372b28673a31 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.073808] env[61768]: DEBUG nova.compute.provider_tree [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1039.249457] env[61768]: DEBUG nova.network.neutron [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.445791] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 8.542s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.485339] env[61768]: DEBUG nova.compute.manager [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-vif-plugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.485560] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.485772] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.485949] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.486139] env[61768]: DEBUG nova.compute.manager [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] No waiting events found dispatching network-vif-plugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1039.486312] env[61768]: WARNING nova.compute.manager [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received unexpected event network-vif-plugged-8e676cc0-ee67-4e76-afe1-2dff2e254617 for instance with vm_state shelved_offloaded and task_state spawning. [ 1039.486478] env[61768]: DEBUG nova.compute.manager [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.486638] env[61768]: DEBUG nova.compute.manager [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing instance network info cache due to event network-changed-8e676cc0-ee67-4e76-afe1-2dff2e254617. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.486817] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.509888] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229464, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.458339} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.510196] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/c4a4d1d5-d91f-4513-9dff-c0a5398ead5e.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1039.510382] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1039.510621] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9524bbcc-69f2-4584-910d-bcd1cefeca3f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.517020] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1039.517020] env[61768]: value = "task-1229465" [ 1039.517020] env[61768]: _type = "Task" [ 1039.517020] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.524980] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229465, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.577010] env[61768]: DEBUG nova.scheduler.client.report [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.745755] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "86f41ef0-4c92-4626-bb83-1148144c597c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.746249] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.746298] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "86f41ef0-4c92-4626-bb83-1148144c597c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.746481] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.746658] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.748868] env[61768]: INFO nova.compute.manager [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Terminating instance [ 1039.750695] env[61768]: DEBUG nova.compute.manager [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1039.750928] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1039.751419] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.753816] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdfbf7c1-4ccd-42c7-84bd-7232592ff793 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.756984] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.757189] env[61768]: DEBUG nova.network.neutron [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Refreshing network info cache for port 8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1039.762853] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1039.763040] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b01d2e4f-c30e-44e1-ba7b-7da72fa4b6d3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.769744] env[61768]: DEBUG oslo_vmware.api [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1039.769744] env[61768]: value = "task-1229466" [ 1039.769744] env[61768]: _type = "Task" [ 1039.769744] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.780042] env[61768]: DEBUG oslo_vmware.api [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229466, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.786387] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e0fc315ed782b9159ff6b38effadc594',container_format='bare',created_at=2024-10-09T20:00:50Z,direct_url=,disk_format='vmdk',id=9b2e869d-9e9c-4311-aae2-355101bc801a,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-169511451-shelved',owner='aa61ecc36bc14769a2fc909ee95716ae',properties=ImageMetaProps,protected=,size=31668736,status='active',tags=,updated_at=2024-10-09T20:01:05Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1039.786654] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1039.786839] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1039.787052] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1039.787215] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1039.787372] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1039.787585] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1039.787753] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1039.787932] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1039.788115] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1039.788298] env[61768]: DEBUG nova.virt.hardware [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1039.789448] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cdf2eeb-de5c-4685-9a5f-704bbe041c12 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.796877] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26154c07-5300-4896-9ac8-4f0c026c3d02 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.810434] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:30:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '37434b93-dfdc-4a3f-bf5a-9f2cbe25a754', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8e676cc0-ee67-4e76-afe1-2dff2e254617', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1039.817783] env[61768]: DEBUG oslo.service.loopingcall [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1039.817909] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1039.818133] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8f4e2970-30d1-470c-a118-01a6986718c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.837098] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1039.837098] env[61768]: value = "task-1229467" [ 1039.837098] env[61768]: _type = "Task" [ 1039.837098] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.845072] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229467, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.026539] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229465, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059921} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.026860] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1040.027609] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37634815-1c0a-4c67-80fc-c44db4ebcbd9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.051822] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/c4a4d1d5-d91f-4513-9dff-c0a5398ead5e.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1040.052204] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c8f74e8-958b-4113-bdc0-86c7012f33d5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.073207] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1040.073207] env[61768]: value = "task-1229468" [ 1040.073207] env[61768]: _type = "Task" [ 1040.073207] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.082643] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.700s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.085165] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229468, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.085617] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.898s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.085831] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.085997] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1040.086926] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d99606-90f7-46cd-9592-0c84de7ff7ac {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.094845] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b9e0d8-e51b-4187-b0ac-1e0f28e04775 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.110619] env[61768]: INFO nova.scheduler.client.report [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Deleted allocations for instance 0e598e1b-754d-4a2f-92e6-2e43463d9e26 [ 1040.112299] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3458d6c1-9548-46be-90f6-26e0b772f615 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.122659] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67adc293-da21-4d19-94bf-07ffc4b89f16 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.153803] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180587MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1040.154150] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.154230] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.279246] env[61768]: DEBUG oslo_vmware.api [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229466, 'name': PowerOffVM_Task, 'duration_secs': 0.37122} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.279541] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1040.279743] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1040.280042] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37face96-7355-4593-a6b5-aa8d5f6d6cfc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.348581] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229467, 'name': CreateVM_Task} progress is 99%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.426893] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1040.427294] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1040.427504] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleting the datastore file [datastore1] 86f41ef0-4c92-4626-bb83-1148144c597c {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1040.428893] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-934274a9-bf9f-40f8-83a9-a24f7f0b87d8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.438592] env[61768]: DEBUG oslo_vmware.api [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1040.438592] env[61768]: value = "task-1229470" [ 1040.438592] env[61768]: _type = "Task" [ 1040.438592] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.446710] env[61768]: DEBUG oslo_vmware.api [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229470, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.528747] env[61768]: DEBUG nova.network.neutron [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updated VIF entry in instance network info cache for port 8e676cc0-ee67-4e76-afe1-2dff2e254617. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.529186] env[61768]: DEBUG nova.network.neutron [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.583165] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229468, 'name': ReconfigVM_Task, 'duration_secs': 0.286741} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.583443] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfigured VM instance instance-0000006e to attach disk [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/c4a4d1d5-d91f-4513-9dff-c0a5398ead5e.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.584113] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a703278d-b561-4c27-9e70-68efc91fe707 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.590408] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1040.590408] env[61768]: value = "task-1229471" [ 1040.590408] env[61768]: _type = "Task" [ 1040.590408] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.606023] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229471, 'name': Rename_Task} progress is 10%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.621342] env[61768]: DEBUG oslo_concurrency.lockutils [None req-968be8f9-d3c8-4bbc-9ebb-07abb78632f7 tempest-ServerShowV257Test-1601172911 tempest-ServerShowV257Test-1601172911-project-member] Lock "0e598e1b-754d-4a2f-92e6-2e43463d9e26" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.858s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.847645] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229467, 'name': CreateVM_Task, 'duration_secs': 0.51451} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.848024] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1040.848453] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.848637] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.849070] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.849337] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-108c937c-9f39-4440-9046-7825ef90c764 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.853767] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1040.853767] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5294d3c3-b5b3-e77b-f5e5-99cc33f9a72c" [ 1040.853767] env[61768]: _type = "Task" [ 1040.853767] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.860838] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]5294d3c3-b5b3-e77b-f5e5-99cc33f9a72c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.936567] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "a96d4721-83c1-474f-8470-a34262e34fad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.936821] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.937047] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "a96d4721-83c1-474f-8470-a34262e34fad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.937250] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.938152] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.942403] env[61768]: INFO nova.compute.manager [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Terminating instance [ 1040.944453] env[61768]: DEBUG nova.compute.manager [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1040.944670] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1040.945464] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcdafeb1-eaa7-4333-a6aa-b059c62b173f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.951662] env[61768]: DEBUG oslo_vmware.api [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229470, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197799} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.952257] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1040.952445] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1040.952621] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1040.952799] env[61768]: INFO nova.compute.manager [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Took 1.20 seconds to destroy the instance on the hypervisor. [ 1040.953092] env[61768]: DEBUG oslo.service.loopingcall [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.953298] env[61768]: DEBUG nova.compute.manager [-] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1040.953394] env[61768]: DEBUG nova.network.neutron [-] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1040.956885] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1040.958106] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82ba202b-fee3-4535-9860-27294014945c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.963236] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1040.963236] env[61768]: value = "task-1229472" [ 1040.963236] env[61768]: _type = "Task" [ 1040.963236] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.972335] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229472, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.032594] env[61768]: DEBUG oslo_concurrency.lockutils [req-cb6bbbd8-092e-4a07-81e7-e601e85b7438 req-07daffed-9ea0-43c0-9dd8-2272ee913bc7 service nova] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.101595] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229471, 'name': Rename_Task, 'duration_secs': 0.141325} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.101987] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1041.102331] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c569ed6d-e1e5-43a9-8e08-4d6ecfa45370 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.109526] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1041.109526] env[61768]: value = "task-1229473" [ 1041.109526] env[61768]: _type = "Task" [ 1041.109526] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.120070] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.180642] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 86f41ef0-4c92-4626-bb83-1148144c597c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.180642] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance e8095d6d-18c5-4ca9-8ac8-475aeeffc01a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.180642] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance aa568a72-afba-4a25-a048-ca62191f9212 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.180642] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance a96d4721-83c1-474f-8470-a34262e34fad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.180897] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 81641ce1-cc88-4df1-a54c-48f9bd4bb73b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.180897] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance c4a4d1d5-d91f-4513-9dff-c0a5398ead5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1041.181093] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1041.181242] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1041.271993] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98a5f97-fcd3-4b33-b407-6c05c87bd5ff {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.281293] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fa8bc90-01cb-42fd-ab0d-5df3940b52ce {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.311546] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9873bb7-a248-4a66-ab7b-7eb2016c0932 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.318945] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bfefd97-3c96-4819-945b-6ed0c134e19c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.332183] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1041.363383] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.363735] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Processing image 9b2e869d-9e9c-4311-aae2-355101bc801a {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1041.364837] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a/9b2e869d-9e9c-4311-aae2-355101bc801a.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.364837] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a/9b2e869d-9e9c-4311-aae2-355101bc801a.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.364837] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1041.364837] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-402c6715-4bbd-4040-9ad4-6bef800bae36 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.375249] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1041.375454] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1041.376215] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be832095-99c2-4dc6-a48d-563761ff3252 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.381542] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1041.381542] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a16ee6-408c-60b9-592e-a75615d87cfb" [ 1041.381542] env[61768]: _type = "Task" [ 1041.381542] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.390246] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52a16ee6-408c-60b9-592e-a75615d87cfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.473222] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229472, 'name': PowerOffVM_Task, 'duration_secs': 0.203973} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.473484] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1041.473654] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1041.473895] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ff07b611-e8e9-4fd0-8c6b-dc7ee2de1b21 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.520926] env[61768]: DEBUG nova.compute.manager [req-07049fda-b50c-420a-ac8d-2bc413150d83 req-51756b98-7199-4544-a0cd-0a9e20e8b12c service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Received event network-vif-deleted-09109871-16a8-42aa-be6d-e34e9885e21d {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1041.521156] env[61768]: INFO nova.compute.manager [req-07049fda-b50c-420a-ac8d-2bc413150d83 req-51756b98-7199-4544-a0cd-0a9e20e8b12c service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Neutron deleted interface 09109871-16a8-42aa-be6d-e34e9885e21d; detaching it from the instance and deleting it from the info cache [ 1041.521380] env[61768]: DEBUG nova.network.neutron [req-07049fda-b50c-420a-ac8d-2bc413150d83 req-51756b98-7199-4544-a0cd-0a9e20e8b12c service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.534383] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1041.534601] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1041.534891] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleting the datastore file [datastore2] a96d4721-83c1-474f-8470-a34262e34fad {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.535386] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5fc9c2f8-6da0-491d-87f7-f955ec61792e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.542128] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for the task: (returnval){ [ 1041.542128] env[61768]: value = "task-1229475" [ 1041.542128] env[61768]: _type = "Task" [ 1041.542128] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.549551] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229475, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.623470] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229473, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.834911] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1041.870286] env[61768]: DEBUG nova.network.neutron [-] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.893157] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Preparing fetch location {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1041.893509] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Fetch image to [datastore2] OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd/OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd.vmdk {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1041.893509] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Downloading stream optimized image 9b2e869d-9e9c-4311-aae2-355101bc801a to [datastore2] OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd/OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd.vmdk on the data store datastore2 as vApp {{(pid=61768) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1041.893668] env[61768]: DEBUG nova.virt.vmwareapi.images [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Downloading image file data 9b2e869d-9e9c-4311-aae2-355101bc801a to the ESX as VM named 'OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd' {{(pid=61768) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1041.969600] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1041.969600] env[61768]: value = "resgroup-9" [ 1041.969600] env[61768]: _type = "ResourcePool" [ 1041.969600] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1041.970013] env[61768]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8ea92145-1d8c-4a0a-8292-0d6c66dbbdbb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.992205] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease: (returnval){ [ 1041.992205] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520a4aaa-bbc7-1407-0f75-60efacc5a86f" [ 1041.992205] env[61768]: _type = "HttpNfcLease" [ 1041.992205] env[61768]: } obtained for vApp import into resource pool (val){ [ 1041.992205] env[61768]: value = "resgroup-9" [ 1041.992205] env[61768]: _type = "ResourcePool" [ 1041.992205] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1041.992493] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the lease: (returnval){ [ 1041.992493] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520a4aaa-bbc7-1407-0f75-60efacc5a86f" [ 1041.992493] env[61768]: _type = "HttpNfcLease" [ 1041.992493] env[61768]: } to be ready. {{(pid=61768) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1041.998969] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1041.998969] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520a4aaa-bbc7-1407-0f75-60efacc5a86f" [ 1041.998969] env[61768]: _type = "HttpNfcLease" [ 1041.998969] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1042.023844] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e196b550-3b90-4f8a-bda8-d75907f657a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.032775] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52e308c0-2e07-4f9f-8ac9-68ce86bec715 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.052723] env[61768]: DEBUG oslo_vmware.api [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Task: {'id': task-1229475, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140979} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.053015] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.053296] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1042.053505] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1042.053702] env[61768]: INFO nova.compute.manager [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1042.053969] env[61768]: DEBUG oslo.service.loopingcall [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.054278] env[61768]: DEBUG nova.compute.manager [-] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.054471] env[61768]: DEBUG nova.network.neutron [-] [instance: a96d4721-83c1-474f-8470-a34262e34fad] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1042.064490] env[61768]: DEBUG nova.compute.manager [req-07049fda-b50c-420a-ac8d-2bc413150d83 req-51756b98-7199-4544-a0cd-0a9e20e8b12c service nova] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Detach interface failed, port_id=09109871-16a8-42aa-be6d-e34e9885e21d, reason: Instance 86f41ef0-4c92-4626-bb83-1148144c597c could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1042.124403] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229473, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.341550] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1042.341815] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.188s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.372926] env[61768]: INFO nova.compute.manager [-] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Took 1.42 seconds to deallocate network for instance. [ 1042.500645] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1042.500645] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520a4aaa-bbc7-1407-0f75-60efacc5a86f" [ 1042.500645] env[61768]: _type = "HttpNfcLease" [ 1042.500645] env[61768]: } is initializing. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1042.623620] env[61768]: DEBUG oslo_vmware.api [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229473, 'name': PowerOnVM_Task, 'duration_secs': 1.02601} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.623968] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1042.624238] env[61768]: INFO nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Took 7.64 seconds to spawn the instance on the hypervisor. [ 1042.624464] env[61768]: DEBUG nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.625585] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-495f8ae7-aa7b-4fc3-bfc6-ac924a7c527a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.821767] env[61768]: DEBUG nova.network.neutron [-] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.879544] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.879841] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.880045] env[61768]: DEBUG nova.objects.instance [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'resources' on Instance uuid 86f41ef0-4c92-4626-bb83-1148144c597c {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.000647] env[61768]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1043.000647] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520a4aaa-bbc7-1407-0f75-60efacc5a86f" [ 1043.000647] env[61768]: _type = "HttpNfcLease" [ 1043.000647] env[61768]: } is ready. {{(pid=61768) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1043.000943] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1043.000943] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]520a4aaa-bbc7-1407-0f75-60efacc5a86f" [ 1043.000943] env[61768]: _type = "HttpNfcLease" [ 1043.000943] env[61768]: }. {{(pid=61768) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1043.001697] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c803bad-bb37-4567-8447-69541cf5bf97 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.008840] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527f31e5-d178-39d7-68da-f24f7874ce10/disk-0.vmdk from lease info. {{(pid=61768) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1043.009044] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating HTTP connection to write to file with size = 31668736 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527f31e5-d178-39d7-68da-f24f7874ce10/disk-0.vmdk. {{(pid=61768) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1043.078539] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-20ab3e5f-b523-4b50-9bb1-78740e0a3e75 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.143028] env[61768]: INFO nova.compute.manager [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Took 13.11 seconds to build instance. [ 1043.325182] env[61768]: INFO nova.compute.manager [-] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Took 1.27 seconds to deallocate network for instance. [ 1043.483295] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d61db19-aa42-46bb-8ba5-05be86d6c88d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.493568] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6feb2e7b-b0f3-4054-8393-4d8c240b9022 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.524840] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9b9939-aed1-4ba2-afa5-a7253db341c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.532916] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ece73b2-8478-48f2-9b1a-c6510046239d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.548682] env[61768]: DEBUG nova.compute.provider_tree [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1043.554277] env[61768]: DEBUG nova.compute.manager [req-4289b6da-825f-41f8-abe0-505766a02d82 req-802ac996-4710-4008-8899-2df647d8c6e6 service nova] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Received event network-vif-deleted-3c8e6155-6935-4025-a622-f7e4f08a0ec3 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1043.648222] env[61768]: DEBUG oslo_concurrency.lockutils [None req-dd12df91-a685-40d1-b1d8-ec7f702bf7a8 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.624s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.834512] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.056357] env[61768]: DEBUG nova.scheduler.client.report [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1044.175520] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Completed reading data from the image iterator. {{(pid=61768) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1044.175840] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527f31e5-d178-39d7-68da-f24f7874ce10/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1044.176949] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7af078-d07e-4aba-8b75-f7b7ad257f6e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.184048] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527f31e5-d178-39d7-68da-f24f7874ce10/disk-0.vmdk is in state: ready. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1044.184224] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527f31e5-d178-39d7-68da-f24f7874ce10/disk-0.vmdk. {{(pid=61768) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1044.184465] env[61768]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-59886f0c-f585-4451-9866-74515cf3cfed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.356577] env[61768]: DEBUG oslo_vmware.rw_handles [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527f31e5-d178-39d7-68da-f24f7874ce10/disk-0.vmdk. {{(pid=61768) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1044.356577] env[61768]: INFO nova.virt.vmwareapi.images [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Downloaded image file data 9b2e869d-9e9c-4311-aae2-355101bc801a [ 1044.357561] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d62d5e75-14da-475f-8414-04efabb5ec9b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.373553] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-273a4887-3fd4-4083-963b-7400abeea8cf {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.398434] env[61768]: INFO nova.virt.vmwareapi.images [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] The imported VM was unregistered [ 1044.400988] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Caching image {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1044.401243] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Creating directory with path [datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1044.401524] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-548ef30b-9147-49ed-802e-e3d6177d7ce2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.411454] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Created directory with path [datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1044.411724] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd/OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd.vmdk to [datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a/9b2e869d-9e9c-4311-aae2-355101bc801a.vmdk. {{(pid=61768) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1044.411885] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-51dfcaaa-af5b-41c8-9301-eb79de9bae55 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.417929] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1044.417929] env[61768]: value = "task-1229478" [ 1044.417929] env[61768]: _type = "Task" [ 1044.417929] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.425472] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229478, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.561949] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.564598] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.730s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.564881] env[61768]: DEBUG nova.objects.instance [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lazy-loading 'resources' on Instance uuid a96d4721-83c1-474f-8470-a34262e34fad {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1044.585032] env[61768]: INFO nova.scheduler.client.report [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted allocations for instance 86f41ef0-4c92-4626-bb83-1148144c597c [ 1044.928713] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229478, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.095747] env[61768]: DEBUG oslo_concurrency.lockutils [None req-1ebdb5fb-3f2f-44ab-bc0d-4c40157700fb tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "86f41ef0-4c92-4626-bb83-1148144c597c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.350s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.165620] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07e24994-5ea5-483d-b2c6-92d90c630ccd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.174743] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02dfde5-0187-4b9c-be0f-75be94c1abdc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.211328] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bad6101-c2ff-4a16-acad-affb5582ab2d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.219561] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c97a7a9-5c25-4bbc-9142-f5fea4f868fd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.234814] env[61768]: DEBUG nova.compute.provider_tree [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1045.428674] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229478, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.580720] env[61768]: DEBUG nova.compute.manager [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Received event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.580868] env[61768]: DEBUG nova.compute.manager [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing instance network info cache due to event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1045.581109] env[61768]: DEBUG oslo_concurrency.lockutils [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] Acquiring lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1045.581294] env[61768]: DEBUG oslo_concurrency.lockutils [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] Acquired lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1045.581466] env[61768]: DEBUG nova.network.neutron [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1045.737608] env[61768]: DEBUG nova.scheduler.client.report [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.929462] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229478, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.242312] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.678s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.266858] env[61768]: INFO nova.scheduler.client.report [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Deleted allocations for instance a96d4721-83c1-474f-8470-a34262e34fad [ 1046.417533] env[61768]: DEBUG nova.network.neutron [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updated VIF entry in instance network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1046.417533] env[61768]: DEBUG nova.network.neutron [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.430305] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229478, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.615448] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1046.782145] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0b3c7c06-4e19-4424-bba3-c119346f81ac tempest-DeleteServersTestJSON-409350480 tempest-DeleteServersTestJSON-409350480-project-member] Lock "a96d4721-83c1-474f-8470-a34262e34fad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.845s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.907953] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.908191] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.919988] env[61768]: DEBUG oslo_concurrency.lockutils [req-8b25466d-4c30-4b91-845f-23cbcd50a8d9 req-21fce992-1b97-4634-91ff-d3f01e6c306e service nova] Releasing lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1046.930802] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229478, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.304439} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.931065] env[61768]: INFO nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd/OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd.vmdk to [datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a/9b2e869d-9e9c-4311-aae2-355101bc801a.vmdk. [ 1046.931263] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Cleaning up location [datastore2] OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1046.931433] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_52f34677-19b8-44d3-b567-cef0e6e46bbd {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1046.931686] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cb934d28-43b7-4db5-93ab-08f43a1e3b58 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.937772] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1046.937772] env[61768]: value = "task-1229479" [ 1046.937772] env[61768]: _type = "Task" [ 1046.937772] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.944868] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229479, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.412378] env[61768]: DEBUG nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Starting instance... {{(pid=61768) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1047.447649] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229479, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.037187} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.447905] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1047.448093] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a/9b2e869d-9e9c-4311-aae2-355101bc801a.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.448350] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a/9b2e869d-9e9c-4311-aae2-355101bc801a.vmdk to [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1047.448610] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-610a64e5-fa35-4510-9ba6-8375752f9495 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.454749] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1047.454749] env[61768]: value = "task-1229481" [ 1047.454749] env[61768]: _type = "Task" [ 1047.454749] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.463382] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229481, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.937458] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.937758] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.939496] env[61768]: INFO nova.compute.claims [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1047.966024] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229481, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.117337] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1048.117545] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1048.117763] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1048.117893] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Cleaning up deleted instances {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1048.466610] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229481, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.634477] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] There are 59 instances to clean {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1048.634771] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 0e598e1b-754d-4a2f-92e6-2e43463d9e26] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1048.967850] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229481, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.029999] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0836911-63f6-40f6-bb85-89c45b3da8ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.037914] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efe583c-c91c-48c3-a3cb-196ade5ae6b8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.069645] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f4b09af-3055-4848-860f-02932500ca4b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.077535] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbadd32-8d35-4fa0-912d-0e6c0e330b61 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.091050] env[61768]: DEBUG nova.compute.provider_tree [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.138102] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: a96d4721-83c1-474f-8470-a34262e34fad] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1049.468024] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229481, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.594681] env[61768]: DEBUG nova.scheduler.client.report [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.641123] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 5bd85959-cb58-4e63-ac1e-7a3ead11cc4f] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1049.971058] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229481, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.228828} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.971058] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9b2e869d-9e9c-4311-aae2-355101bc801a/9b2e869d-9e9c-4311-aae2-355101bc801a.vmdk to [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1049.971317] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32290eec-dafa-4de7-bea1-0957d1ca0453 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.996610] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk or device None with type streamOptimized {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1049.996962] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2085b0a-696d-47d3-9f70-833f99a42ff9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.019075] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1050.019075] env[61768]: value = "task-1229482" [ 1050.019075] env[61768]: _type = "Task" [ 1050.019075] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.026706] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229482, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.099910] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.162s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.100440] env[61768]: DEBUG nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Start building networks asynchronously for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1050.144783] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e0a19f6b-43e7-4a77-9d33-3dfc829bf0aa] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1050.529411] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229482, 'name': ReconfigVM_Task, 'duration_secs': 0.291745} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.529767] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b/81641ce1-cc88-4df1-a54c-48f9bd4bb73b.vmdk or device None with type streamOptimized {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1050.530363] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c107010b-1ecc-4e6a-a371-715a3891bf3e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.536153] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1050.536153] env[61768]: value = "task-1229483" [ 1050.536153] env[61768]: _type = "Task" [ 1050.536153] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.543343] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229483, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.606102] env[61768]: DEBUG nova.compute.utils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1050.607520] env[61768]: DEBUG nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Allocating IP information in the background. {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1050.607688] env[61768]: DEBUG nova.network.neutron [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] allocate_for_instance() {{(pid=61768) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1050.647910] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 86f41ef0-4c92-4626-bb83-1148144c597c] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1050.651244] env[61768]: DEBUG nova.policy [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2280e9e6bde740c3aca30ebf70c7f73d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fc444d04ff864300a45a55a3c67507e5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61768) authorize /opt/stack/nova/nova/policy.py:201}} [ 1050.976311] env[61768]: DEBUG nova.network.neutron [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Successfully created port: 8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1051.047783] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229483, 'name': Rename_Task, 'duration_secs': 0.141055} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.048096] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1051.048357] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7b250fc-84fb-4700-9f41-97cf47994995 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.055152] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1051.055152] env[61768]: value = "task-1229484" [ 1051.055152] env[61768]: _type = "Task" [ 1051.055152] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.063708] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229484, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.110727] env[61768]: DEBUG nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Start building block device mappings for instance. {{(pid=61768) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1051.153840] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 9bbacc8d-e89a-456e-9bbe-871c2a9d4c4f] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1051.565612] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229484, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.657609] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 750bc93c-65aa-4afd-a07b-cbeda0acae24] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1052.066689] env[61768]: DEBUG oslo_vmware.api [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229484, 'name': PowerOnVM_Task, 'duration_secs': 0.537353} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.067111] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1052.125396] env[61768]: DEBUG nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Start spawning the instance on the hypervisor. {{(pid=61768) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1052.150027] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-09T19:50:40Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-09T19:50:23Z,direct_url=,disk_format='vmdk',id=0a07405a-3ac2-4d7f-bd39-a1b659ab2354,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='cf94ddb28c154591afc09dceed563fad',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-09T19:50:23Z,virtual_size=,visibility=), allow threads: False {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1052.150476] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1052.150932] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image limits 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1052.151233] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Flavor pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1052.151315] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Image pref 0:0:0 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1052.151502] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61768) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1052.151891] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1052.152009] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1052.152229] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Got 1 possible topologies {{(pid=61768) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1052.152479] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1052.152845] env[61768]: DEBUG nova.virt.hardware [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61768) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1052.153764] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943b9964-38d8-4389-bed4-f17b839cd2a4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.162190] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17253ed4-322c-4493-979c-7361565b6ee4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.166403] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 49723fff-d1ce-476f-ac23-bd2472762d94] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1052.179983] env[61768]: DEBUG nova.compute.manager [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.180810] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea7eb753-e6c9-43d5-b8ff-e6b04cd51625 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.411804] env[61768]: DEBUG nova.compute.manager [req-5da89d1d-37bf-45c0-8589-a2834d95731a req-44c211b6-3f38-4d27-b727-1bfc1b0a6967 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Received event network-vif-plugged-8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.412086] env[61768]: DEBUG oslo_concurrency.lockutils [req-5da89d1d-37bf-45c0-8589-a2834d95731a req-44c211b6-3f38-4d27-b727-1bfc1b0a6967 service nova] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1052.412325] env[61768]: DEBUG oslo_concurrency.lockutils [req-5da89d1d-37bf-45c0-8589-a2834d95731a req-44c211b6-3f38-4d27-b727-1bfc1b0a6967 service nova] Lock "4d34dea4-917e-41a3-ac7c-706e97603711-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.412498] env[61768]: DEBUG oslo_concurrency.lockutils [req-5da89d1d-37bf-45c0-8589-a2834d95731a req-44c211b6-3f38-4d27-b727-1bfc1b0a6967 service nova] Lock "4d34dea4-917e-41a3-ac7c-706e97603711-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.412682] env[61768]: DEBUG nova.compute.manager [req-5da89d1d-37bf-45c0-8589-a2834d95731a req-44c211b6-3f38-4d27-b727-1bfc1b0a6967 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] No waiting events found dispatching network-vif-plugged-8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1052.412871] env[61768]: WARNING nova.compute.manager [req-5da89d1d-37bf-45c0-8589-a2834d95731a req-44c211b6-3f38-4d27-b727-1bfc1b0a6967 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Received unexpected event network-vif-plugged-8ec9169c-da5b-4f41-9bd9-76d42ac011e6 for instance with vm_state building and task_state spawning. [ 1052.506126] env[61768]: DEBUG nova.network.neutron [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Successfully updated port: 8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1052.669504] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 271fe080-b8e9-4141-b2f3-4c4efaad81c7] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1052.694621] env[61768]: DEBUG oslo_concurrency.lockutils [None req-961afe4b-d256-4ccc-b3f6-3ffda9a19eda tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 23.511s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.008915] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.009172] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.009266] env[61768]: DEBUG nova.network.neutron [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1053.173400] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e1e90c98-5a65-473c-b508-b45fd93e31ad] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1053.540888] env[61768]: DEBUG nova.network.neutron [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Instance cache missing network info. {{(pid=61768) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1053.661504] env[61768]: DEBUG nova.network.neutron [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updating instance_info_cache with network_info: [{"id": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "address": "fa:16:3e:69:96:48", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec9169c-da", "ovs_interfaceid": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.677238] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 71ce99fa-3d9d-4ff6-b06a-2e27d575ab98] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1054.164752] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.165071] env[61768]: DEBUG nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Instance network_info: |[{"id": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "address": "fa:16:3e:69:96:48", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec9169c-da", "ovs_interfaceid": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61768) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1054.165578] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:69:96:48', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13e71dbb-4279-427c-b39d-ba5df9895e58', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ec9169c-da5b-4f41-9bd9-76d42ac011e6', 'vif_model': 'vmxnet3'}] {{(pid=61768) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1054.173134] env[61768]: DEBUG oslo.service.loopingcall [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1054.173350] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Creating VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1054.173572] env[61768]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-829a6567-baf6-4f8a-a38e-da3fa65618b4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.188518] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: df570987-0144-4e3a-83db-a9538d3835f5] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1054.196479] env[61768]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1054.196479] env[61768]: value = "task-1229485" [ 1054.196479] env[61768]: _type = "Task" [ 1054.196479] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.204204] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229485, 'name': CreateVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.441853] env[61768]: DEBUG nova.compute.manager [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Received event network-changed-8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.441948] env[61768]: DEBUG nova.compute.manager [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Refreshing instance network info cache due to event network-changed-8ec9169c-da5b-4f41-9bd9-76d42ac011e6. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1054.442165] env[61768]: DEBUG oslo_concurrency.lockutils [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] Acquiring lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.442330] env[61768]: DEBUG oslo_concurrency.lockutils [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] Acquired lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.442506] env[61768]: DEBUG nova.network.neutron [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Refreshing network info cache for port 8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1054.692176] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 1158fbc3-f17b-44f7-847f-01bdf328a74a] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1054.706773] env[61768]: DEBUG oslo_vmware.api [-] Task: {'id': task-1229485, 'name': CreateVM_Task, 'duration_secs': 0.326507} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.706936] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Created VM on the ESX host {{(pid=61768) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1054.707670] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.707873] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.708215] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1054.708491] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f46ec1af-3c50-4f61-b37c-cb6d24cc175c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.713952] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1054.713952] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529e1b87-9d0f-d986-4eef-2eda62490812" [ 1054.713952] env[61768]: _type = "Task" [ 1054.713952] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.722323] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529e1b87-9d0f-d986-4eef-2eda62490812, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.165390] env[61768]: DEBUG nova.network.neutron [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updated VIF entry in instance network info cache for port 8ec9169c-da5b-4f41-9bd9-76d42ac011e6. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1055.165774] env[61768]: DEBUG nova.network.neutron [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updating instance_info_cache with network_info: [{"id": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "address": "fa:16:3e:69:96:48", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec9169c-da", "ovs_interfaceid": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.195788] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: fb344f81-9295-44a1-9bac-dd530c157e98] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1055.224749] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]529e1b87-9d0f-d986-4eef-2eda62490812, 'name': SearchDatastore_Task, 'duration_secs': 0.041887} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.225549] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.225791] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1055.226143] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1055.226330] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.226531] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1055.226998] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-788eee65-39c0-4e91-9c2e-c71b1704386b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.242522] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1055.242784] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1055.243567] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af803a18-16d9-4ab3-a58a-237452289dcb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.249506] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1055.249506] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525ee142-e896-2032-5a8f-6ffd02c9c885" [ 1055.249506] env[61768]: _type = "Task" [ 1055.249506] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.256842] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525ee142-e896-2032-5a8f-6ffd02c9c885, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.669078] env[61768]: DEBUG oslo_concurrency.lockutils [req-3d819d28-59fc-4483-ab2d-7fadbdc5e7f0 req-95c46ddc-9adc-4255-aaea-d67d6c132e90 service nova] Releasing lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.698862] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e0c89fa6-9d32-4343-89d7-e9129d252b40] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1055.760280] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]525ee142-e896-2032-5a8f-6ffd02c9c885, 'name': SearchDatastore_Task, 'duration_secs': 0.041947} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.761063] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c606a4f1-257a-4a4e-978f-62833b2f62f0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.766058] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1055.766058] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f2fc09-9625-6465-4c7c-4960dbbfac66" [ 1055.766058] env[61768]: _type = "Task" [ 1055.766058] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.773343] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f2fc09-9625-6465-4c7c-4960dbbfac66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.202494] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 30f55051-63b3-47a9-83fb-945cdeb82574] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1056.277667] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52f2fc09-9625-6465-4c7c-4960dbbfac66, 'name': SearchDatastore_Task, 'duration_secs': 0.009594} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.277952] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.278298] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 4d34dea4-917e-41a3-ac7c-706e97603711/4d34dea4-917e-41a3-ac7c-706e97603711.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1056.278614] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf864bd1-d941-4d65-99f0-6cddcd19d57e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.286037] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1056.286037] env[61768]: value = "task-1229486" [ 1056.286037] env[61768]: _type = "Task" [ 1056.286037] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.294177] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.706828] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 739faae0-5811-4f88-b56b-1350c0f7b8be] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1056.795062] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.457697} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.795342] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] 4d34dea4-917e-41a3-ac7c-706e97603711/4d34dea4-917e-41a3-ac7c-706e97603711.vmdk {{(pid=61768) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1056.795590] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Extending root virtual disk to 1048576 {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1056.796097] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-199bfe09-6606-40a0-83c9-85e217a9cb5e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.802992] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1056.802992] env[61768]: value = "task-1229487" [ 1056.802992] env[61768]: _type = "Task" [ 1056.802992] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.811540] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229487, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.211066] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: d54c94a6-0a53-47a8-b024-1a05439da837] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1057.312847] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229487, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062287} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.313193] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Extended root virtual disk {{(pid=61768) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1057.313909] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35200a37-17ca-4920-acd3-96924a365d4a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.336177] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] 4d34dea4-917e-41a3-ac7c-706e97603711/4d34dea4-917e-41a3-ac7c-706e97603711.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1057.336454] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b9b1157e-6491-450d-beaf-55e76a5d92e0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.357016] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1057.357016] env[61768]: value = "task-1229488" [ 1057.357016] env[61768]: _type = "Task" [ 1057.357016] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.365634] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229488, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.714250] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: aaf205c5-dd11-4d1c-90bd-3ecd5a121227] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1057.867055] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229488, 'name': ReconfigVM_Task, 'duration_secs': 0.26664} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.867377] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Reconfigured VM instance instance-0000006f to attach disk [datastore2] 4d34dea4-917e-41a3-ac7c-706e97603711/4d34dea4-917e-41a3-ac7c-706e97603711.vmdk or device None with type sparse {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1057.868012] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-263c35ed-e39d-4baa-b8b9-ab11849d8e03 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.874677] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1057.874677] env[61768]: value = "task-1229489" [ 1057.874677] env[61768]: _type = "Task" [ 1057.874677] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.881970] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229489, 'name': Rename_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.217410] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e2b7cb22-475b-4b1b-b30e-6061e888691e] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1058.385119] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229489, 'name': Rename_Task, 'duration_secs': 0.13322} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.385411] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1058.385655] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90942ba2-c58f-4576-a948-012b225f1afc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.392392] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1058.392392] env[61768]: value = "task-1229490" [ 1058.392392] env[61768]: _type = "Task" [ 1058.392392] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.399701] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.720951] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 5cc705a5-1423-43e1-8356-9abd67a8da3a] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1058.901934] env[61768]: DEBUG oslo_vmware.api [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229490, 'name': PowerOnVM_Task, 'duration_secs': 0.422969} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.902232] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1058.902448] env[61768]: INFO nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Took 6.78 seconds to spawn the instance on the hypervisor. [ 1058.902635] env[61768]: DEBUG nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1058.903396] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90ee8442-bacd-43c7-9aba-9bd34a3d6ae0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.224924] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 12f92761-45b6-49c6-96bd-1c0fa05521c6] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1059.419432] env[61768]: INFO nova.compute.manager [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Took 11.50 seconds to build instance. [ 1059.728136] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: c7cc9c7d-a9b9-4d43-b926-72eb1cedd69e] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1059.921607] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e54746a3-c52a-4400-bab0-7ba5486c1a0d tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.013s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.089774] env[61768]: DEBUG nova.compute.manager [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Received event network-changed-8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.090015] env[61768]: DEBUG nova.compute.manager [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Refreshing instance network info cache due to event network-changed-8ec9169c-da5b-4f41-9bd9-76d42ac011e6. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1060.090269] env[61768]: DEBUG oslo_concurrency.lockutils [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] Acquiring lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.090424] env[61768]: DEBUG oslo_concurrency.lockutils [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] Acquired lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.090592] env[61768]: DEBUG nova.network.neutron [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Refreshing network info cache for port 8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1060.231271] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: d13673f2-e99f-400b-bcdc-b7c38e197610] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1060.737036] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: ecd08348-b22f-49c5-b13f-b187506e38d6] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1060.801697] env[61768]: DEBUG nova.network.neutron [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updated VIF entry in instance network info cache for port 8ec9169c-da5b-4f41-9bd9-76d42ac011e6. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1060.802218] env[61768]: DEBUG nova.network.neutron [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updating instance_info_cache with network_info: [{"id": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "address": "fa:16:3e:69:96:48", "network": {"id": "af38316f-9b1c-421b-9d56-fde11088bdf6", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-950541299-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.241", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fc444d04ff864300a45a55a3c67507e5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13e71dbb-4279-427c-b39d-ba5df9895e58", "external-id": "nsx-vlan-transportzone-417", "segmentation_id": 417, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ec9169c-da", "ovs_interfaceid": "8ec9169c-da5b-4f41-9bd9-76d42ac011e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1061.239913] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 86e24eb9-6a55-4658-9e62-64713b0289d7] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1061.305261] env[61768]: DEBUG oslo_concurrency.lockutils [req-3e36bff0-2e6a-42df-8e1f-e3c4c43d5c54 req-06d7c39f-b388-453f-9c70-4aab39f68bfb service nova] Releasing lock "refresh_cache-4d34dea4-917e-41a3-ac7c-706e97603711" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.742903] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: aebe41ad-496d-4a53-b023-c8df2bca04f8] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1062.246512] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 41f6300d-7462-4e11-b32b-8c892e87bafc] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1062.749243] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 8e0e13a7-eac5-4176-8536-2906d13c390e] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1063.252175] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 37b729bf-ad5e-4e17-b11b-77b504c049cd] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1063.755871] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 3aa722e5-0818-40ae-8220-223d920a7386] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.259375] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: d4a765bd-a56d-44b7-8db3-c081832d58c9] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1064.762822] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: b6e4f743-5692-4b46-892c-6c8917ccef98] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.266224] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 356a69b0-7898-440d-9473-7b4572ed2315] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1065.770770] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: d928f14c-b94a-4abf-a053-51015f3bc6c5] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.274628] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 23cc444d-ff15-441b-8a0c-2ded49fd3cf0] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1066.778328] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: a36b542c-334f-46b1-9030-ee5c9f8c55c1] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.281337] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 7697ecda-ef30-447d-a92a-3f5cb3cc9118] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1067.785050] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 4f429be7-c594-4a72-aebd-0b746219231c] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.288253] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 3d96e47d-fc09-439e-bf0e-e60fc87ae958] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1068.791245] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: fbb8c9db-e823-444a-84de-b231cf9b8a8a] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.294082] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 2c85de5d-0111-40a6-a8bc-69c7eba0393d] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1069.797240] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 62cddff5-d499-4ef3-869d-3cdb0328640d] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.300827] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: a428f5e0-ade1-4aa6-af9d-0e33efcfec62] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1070.803921] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 77319462-e447-405b-9269-82581effe005] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.307640] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 76c282d1-ddfe-46dc-aa7b-225708443379] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1071.810662] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 659e8d4c-1af6-4e81-b6a1-2d9466af6b0f] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.314159] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 9e8f6cdd-31fe-4f7b-b13a-4fbaa0d6ce54] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1072.817554] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: d70040f1-e425-4aeb-b82d-ce808a02a645] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.320842] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 23a4450d-162a-4b2b-a009-7023851315a1] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1073.824316] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 9b3632e5-4c2d-4968-9692-70f91883bfb3] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.327261] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 9fa32a4d-3ba4-4d36-963b-17a64453e804] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1074.831094] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 668d4fe4-85f8-4282-8d65-0549d9e3bda8] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.334192] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 7ef50ee4-64e6-49c2-bb41-0bdce316f8d3] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1075.837738] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: efbde802-8cb0-4563-a776-3722a3889afe] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.340984] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 91d56a8c-0664-4b3c-bf28-e2c668ac65fa] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1076.844913] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 0df161e9-4a66-432e-9366-60f231c53e3d] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.348158] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 663fa2fb-1ced-4052-9eba-8a5d0cbe96b5] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1077.851302] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 12a543c2-1081-49d7-800b-07f0a2516904] Instance has had 0 of 5 cleanup attempts {{(pid=61768) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1078.354356] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.354525] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Cleaning up deleted instances with incomplete migration {{(pid=61768) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1079.354693] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.355072] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.860834] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.861077] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1079.861077] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Rebuilding the list of instances to heal {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1080.379233] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.379530] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.392240] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.392379] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquired lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.392525] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Forcefully refreshing network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1080.392681] env[61768]: DEBUG nova.objects.instance [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lazy-loading 'info_cache' on Instance uuid 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1080.882966] env[61768]: DEBUG nova.compute.utils [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1081.385570] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1082.147327] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [{"id": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "address": "fa:16:3e:a5:30:97", "network": {"id": "9af342d6-2fa6-4ad4-b40c-6d255f4fc7d3", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-1578357004-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.174", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa61ecc36bc14769a2fc909ee95716ae", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "37434b93-dfdc-4a3f-bf5a-9f2cbe25a754", "external-id": "nsx-vlan-transportzone-676", "segmentation_id": 676, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8e676cc0-ee", "ovs_interfaceid": "8e676cc0-ee67-4e76-afe1-2dff2e254617", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.447416] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.447791] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.447916] env[61768]: INFO nova.compute.manager [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Attaching volume cbb7e026-7118-4822-b03a-1afcd8b91b04 to /dev/sdb [ 1082.477556] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9a2d15-ca3b-449e-afef-0342db164949 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.484353] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe2f522-dd72-47bc-bf37-7a2d094a81ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.496747] env[61768]: DEBUG nova.virt.block_device [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating existing volume attachment record: 2ce4d30a-76f9-47dd-b0cd-a56878f899b5 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1082.649521] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Releasing lock "refresh_cache-81641ce1-cc88-4df1-a54c-48f9bd4bb73b" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.649752] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updated the network info_cache for instance {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1082.649979] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.650161] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.650315] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.650456] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.650598] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1082.650755] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_power_states {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.155208] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Getting list of instances from cluster (obj){ [ 1083.155208] env[61768]: value = "domain-c8" [ 1083.155208] env[61768]: _type = "ClusterComputeResource" [ 1083.155208] env[61768]: } {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1083.156320] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e22043-34e5-4fef-859f-d25df15b0723 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.171398] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Got total of 5 instances {{(pid=61768) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1083.171558] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Triggering sync for uuid 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1083.171751] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Triggering sync for uuid e8095d6d-18c5-4ca9-8ac8-475aeeffc01a {{(pid=61768) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1083.171959] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Triggering sync for uuid aa568a72-afba-4a25-a048-ca62191f9212 {{(pid=61768) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1083.172139] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Triggering sync for uuid c4a4d1d5-d91f-4513-9dff-c0a5398ead5e {{(pid=61768) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1083.172310] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Triggering sync for uuid 4d34dea4-917e-41a3-ac7c-706e97603711 {{(pid=61768) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1083.172635] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.172863] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.173137] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.173332] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.173566] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "aa568a72-afba-4a25-a048-ca62191f9212" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.173754] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "aa568a72-afba-4a25-a048-ca62191f9212" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.173982] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.174225] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.174405] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.174604] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.176006] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e97eba-5fbe-4289-b8f4-2365fa6592cb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.178866] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0417b8a6-ce43-481d-8466-6f48ad13675f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.181494] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff6f263-36cd-448c-bd86-eab8c6551bed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.184254] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776268f8-aeec-42c0-a049-8d3623c20900 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.187578] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.187777] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.187940] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.188099] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1083.189044] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0a2882-812c-4af1-828d-6f48c6d8f1a1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.206740] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7c05b4-0068-4974-96d2-1d398f7439f2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.222017] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609088c2-442f-4cab-848e-895665411eab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.228485] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ede029-8a73-42ab-bf85-da737675f8c2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.256782] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180425MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1083.257108] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1083.257424] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1083.706550] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.532s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.706971] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.534s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.707270] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "aa568a72-afba-4a25-a048-ca62191f9212" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.533s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.725885] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.553s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.360030] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance e8095d6d-18c5-4ca9-8ac8-475aeeffc01a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1084.360030] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance aa568a72-afba-4a25-a048-ca62191f9212 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1084.360195] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 81641ce1-cc88-4df1-a54c-48f9bd4bb73b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1084.360268] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance c4a4d1d5-d91f-4513-9dff-c0a5398ead5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1084.360417] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 4d34dea4-917e-41a3-ac7c-706e97603711 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1084.360620] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1084.360761] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1084.419910] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-641022eb-107a-4ec9-ad05-098f2537dc63 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.427434] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b6df29-8b2d-4870-a95c-6d5ae17d0fb3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.456168] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d97a9c2f-cc52-4583-81b6-93cd678ed4c3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.462768] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd225bda-f463-4f9d-a23c-172d0b0faa8a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.475100] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1084.978322] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1085.482751] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1085.483019] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.226s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.039147] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1087.039424] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265550', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'name': 'volume-cbb7e026-7118-4822-b03a-1afcd8b91b04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4a4d1d5-d91f-4513-9dff-c0a5398ead5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'serial': 'cbb7e026-7118-4822-b03a-1afcd8b91b04'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1087.040316] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-848f5374-e8ca-4a21-8b3e-e863a4cadb46 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.056863] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-703e9f03-61d7-4d6c-8174-5d494f635fd2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.082073] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] volume-cbb7e026-7118-4822-b03a-1afcd8b91b04/volume-cbb7e026-7118-4822-b03a-1afcd8b91b04.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1087.082329] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-175a5e70-b75a-4430-b0fa-edd908f95cf3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.101305] env[61768]: DEBUG oslo_vmware.api [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1087.101305] env[61768]: value = "task-1229495" [ 1087.101305] env[61768]: _type = "Task" [ 1087.101305] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.108855] env[61768]: DEBUG oslo_vmware.api [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229495, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.610745] env[61768]: DEBUG oslo_vmware.api [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229495, 'name': ReconfigVM_Task, 'duration_secs': 0.344426} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.611049] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfigured VM instance instance-0000006e to attach disk [datastore2] volume-cbb7e026-7118-4822-b03a-1afcd8b91b04/volume-cbb7e026-7118-4822-b03a-1afcd8b91b04.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1087.615660] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3b91dd5-e23c-4297-b205-d026ed36fc1f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.630377] env[61768]: DEBUG oslo_vmware.api [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1087.630377] env[61768]: value = "task-1229496" [ 1087.630377] env[61768]: _type = "Task" [ 1087.630377] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.637845] env[61768]: DEBUG oslo_vmware.api [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229496, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.753676] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.753982] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.754241] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.754438] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.754621] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.758109] env[61768]: INFO nova.compute.manager [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Terminating instance [ 1087.760007] env[61768]: DEBUG nova.compute.manager [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1087.760224] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1087.761042] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-528139d2-1f41-4b32-81bb-7c4ee2e0c0e3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.768218] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1087.768737] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2cdc2e67-6f86-497c-a2f5-d4df4a1ba013 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.774310] env[61768]: DEBUG oslo_vmware.api [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1087.774310] env[61768]: value = "task-1229497" [ 1087.774310] env[61768]: _type = "Task" [ 1087.774310] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.781961] env[61768]: DEBUG oslo_vmware.api [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229497, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.141838] env[61768]: DEBUG oslo_vmware.api [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229496, 'name': ReconfigVM_Task, 'duration_secs': 0.131211} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.142214] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265550', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'name': 'volume-cbb7e026-7118-4822-b03a-1afcd8b91b04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4a4d1d5-d91f-4513-9dff-c0a5398ead5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'serial': 'cbb7e026-7118-4822-b03a-1afcd8b91b04'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1088.284581] env[61768]: DEBUG oslo_vmware.api [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229497, 'name': PowerOffVM_Task, 'duration_secs': 0.194054} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.285115] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1088.285308] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1088.285565] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6214b6d-f070-406a-987a-6b765041f11e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.349914] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1088.350206] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1088.350400] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleting the datastore file [datastore2] 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1088.350671] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13c7b7d5-65bc-41d4-8bc8-fa215c88d1b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.357936] env[61768]: DEBUG oslo_vmware.api [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for the task: (returnval){ [ 1088.357936] env[61768]: value = "task-1229499" [ 1088.357936] env[61768]: _type = "Task" [ 1088.357936] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.369329] env[61768]: DEBUG oslo_vmware.api [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229499, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.868444] env[61768]: DEBUG oslo_vmware.api [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Task: {'id': task-1229499, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.131168} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.868706] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1088.868904] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1088.869102] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1088.869287] env[61768]: INFO nova.compute.manager [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1088.869536] env[61768]: DEBUG oslo.service.loopingcall [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1088.869727] env[61768]: DEBUG nova.compute.manager [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1088.869847] env[61768]: DEBUG nova.network.neutron [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1089.180453] env[61768]: DEBUG nova.objects.instance [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lazy-loading 'flavor' on Instance uuid c4a4d1d5-d91f-4513-9dff-c0a5398ead5e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.286722] env[61768]: DEBUG nova.compute.manager [req-6b653964-31c9-4eb2-a94e-61e2e8db1755 req-8da3e6c1-6fb0-4282-8ba3-2d8fe3dd73d3 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Received event network-vif-deleted-8e676cc0-ee67-4e76-afe1-2dff2e254617 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1089.286908] env[61768]: INFO nova.compute.manager [req-6b653964-31c9-4eb2-a94e-61e2e8db1755 req-8da3e6c1-6fb0-4282-8ba3-2d8fe3dd73d3 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Neutron deleted interface 8e676cc0-ee67-4e76-afe1-2dff2e254617; detaching it from the instance and deleting it from the info cache [ 1089.287113] env[61768]: DEBUG nova.network.neutron [req-6b653964-31c9-4eb2-a94e-61e2e8db1755 req-8da3e6c1-6fb0-4282-8ba3-2d8fe3dd73d3 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.684860] env[61768]: DEBUG oslo_concurrency.lockutils [None req-aee6d31e-2469-484a-9b4e-23aa28a0247e tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.237s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.685844] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.512s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.687118] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b73815f-0a0a-4819-bd2e-8263d5661de3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.764234] env[61768]: DEBUG nova.network.neutron [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.789483] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-92398f80-5ceb-4d22-b1b2-775c9fb05661 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.800229] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9f49fb-b6fd-4340-8fea-8d55fdb47f3f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.826234] env[61768]: DEBUG nova.compute.manager [req-6b653964-31c9-4eb2-a94e-61e2e8db1755 req-8da3e6c1-6fb0-4282-8ba3-2d8fe3dd73d3 service nova] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Detach interface failed, port_id=8e676cc0-ee67-4e76-afe1-2dff2e254617, reason: Instance 81641ce1-cc88-4df1-a54c-48f9bd4bb73b could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1089.910196] env[61768]: INFO nova.compute.manager [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Rescuing [ 1089.910536] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.910699] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.910879] env[61768]: DEBUG nova.network.neutron [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1090.197697] env[61768]: INFO nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] During sync_power_state the instance has a pending task (rescuing). Skip. [ 1090.198059] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.512s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.266767] env[61768]: INFO nova.compute.manager [-] [instance: 81641ce1-cc88-4df1-a54c-48f9bd4bb73b] Took 1.40 seconds to deallocate network for instance. [ 1090.614919] env[61768]: DEBUG nova.network.neutron [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.772745] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.773046] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.773299] env[61768]: DEBUG nova.objects.instance [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lazy-loading 'resources' on Instance uuid 81641ce1-cc88-4df1-a54c-48f9bd4bb73b {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.117701] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.345459] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81495396-df0c-4a13-bddc-2c01a5b416a2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.353231] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f04235a-afe4-40ef-908c-7d89608e1205 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.382097] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba794e3-9b77-474f-9b74-62ae49b57720 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.389287] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10166499-4b56-4f25-8cad-be5f5f227de3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.401940] env[61768]: DEBUG nova.compute.provider_tree [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.650559] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1091.650843] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-220fc446-858d-41de-9377-032304aa63ab {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.658647] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1091.658647] env[61768]: value = "task-1229500" [ 1091.658647] env[61768]: _type = "Task" [ 1091.658647] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.666251] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.905186] env[61768]: DEBUG nova.scheduler.client.report [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1092.169179] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229500, 'name': PowerOffVM_Task, 'duration_secs': 0.181764} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.169477] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1092.170319] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-401344af-e936-4b8d-9130-294b6d31c59c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.193938] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5501578d-1eaf-4a18-bcd6-30fd0d1c85b0 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.219085] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1092.219329] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3ad8d53e-bc18-4244-94ad-605fcf2251c7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.225779] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1092.225779] env[61768]: value = "task-1229501" [ 1092.225779] env[61768]: _type = "Task" [ 1092.225779] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.232712] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229501, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.409889] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.430065] env[61768]: INFO nova.scheduler.client.report [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Deleted allocations for instance 81641ce1-cc88-4df1-a54c-48f9bd4bb73b [ 1092.736235] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] VM already powered off {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1092.736475] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Processing image 0a07405a-3ac2-4d7f-bd39-a1b659ab2354 {{(pid=61768) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1092.736725] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.736878] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.737173] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1092.737329] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-aec0b0bf-4159-4fd0-93d2-647a5bcf4826 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.747224] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61768) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1092.747416] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61768) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1092.748139] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0e3a0c7-0377-4a84-9b8c-842bf5436534 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.753044] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1092.753044] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ffae76-92cc-5617-2211-c3664f0ff77c" [ 1092.753044] env[61768]: _type = "Task" [ 1092.753044] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.760433] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ffae76-92cc-5617-2211-c3664f0ff77c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.938964] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e2f1827d-e126-4832-9907-505d51005cdc tempest-AttachVolumeShelveTestJSON-1276952064 tempest-AttachVolumeShelveTestJSON-1276952064-project-member] Lock "81641ce1-cc88-4df1-a54c-48f9bd4bb73b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.185s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.263440] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52ffae76-92cc-5617-2211-c3664f0ff77c, 'name': SearchDatastore_Task, 'duration_secs': 0.00823} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.264172] env[61768]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cd0937d-b827-4364-a24f-702b3182d5e3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.269260] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1093.269260] env[61768]: value = "session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fae0fe-da4e-61f9-57b9-5433116f9416" [ 1093.269260] env[61768]: _type = "Task" [ 1093.269260] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.276849] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fae0fe-da4e-61f9-57b9-5433116f9416, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.780265] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': session[52ec4e0d-430e-64f1-322c-df7f0d9e648b]52fae0fe-da4e-61f9-57b9-5433116f9416, 'name': SearchDatastore_Task, 'duration_secs': 0.009584} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.780627] env[61768]: DEBUG oslo_concurrency.lockutils [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.780805] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. {{(pid=61768) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1093.781105] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-262fc9a3-c9c5-49e1-ae7c-43e39a505654 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.788180] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1093.788180] env[61768]: value = "task-1229502" [ 1093.788180] env[61768]: _type = "Task" [ 1093.788180] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.795441] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229502, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.298386] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229502, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.420452} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.298681] env[61768]: INFO nova.virt.vmwareapi.ds_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/0a07405a-3ac2-4d7f-bd39-a1b659ab2354/0a07405a-3ac2-4d7f-bd39-a1b659ab2354.vmdk to [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk. [ 1094.299513] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32853c7-4e96-4eff-9d65-bae96c8ff089 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.327047] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1094.327331] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-adf35e7f-ae86-4453-8c83-4edf2b52b737 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.344397] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1094.344397] env[61768]: value = "task-1229504" [ 1094.344397] env[61768]: _type = "Task" [ 1094.344397] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.354060] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229504, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.854205] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229504, 'name': ReconfigVM_Task, 'duration_secs': 0.316571} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.854610] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfigured VM instance instance-0000006e to attach disk [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e/0a07405a-3ac2-4d7f-bd39-a1b659ab2354-rescue.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1094.855343] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8757015-0661-4de2-9893-b939df5a74b1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.881766] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca346ed1-51a7-47d9-8986-0b07d1b90aad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.896414] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1094.896414] env[61768]: value = "task-1229505" [ 1094.896414] env[61768]: _type = "Task" [ 1094.896414] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.903553] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229505, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.406780] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229505, 'name': ReconfigVM_Task, 'duration_secs': 0.189348} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.407060] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1095.407296] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76e3c988-0114-472a-9c58-dff0eb6749a5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.414066] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1095.414066] env[61768]: value = "task-1229506" [ 1095.414066] env[61768]: _type = "Task" [ 1095.414066] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.422255] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229506, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.924026] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229506, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.424972] env[61768]: DEBUG oslo_vmware.api [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229506, 'name': PowerOnVM_Task, 'duration_secs': 0.634801} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.425286] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1096.428176] env[61768]: DEBUG nova.compute.manager [None req-d7b8175f-f840-4abc-8efc-e5fdcab47d98 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1096.428975] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1647d4fc-bf48-4aa6-874e-3bea76487771 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.089729] env[61768]: INFO nova.compute.manager [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Unrescuing [ 1098.090095] env[61768]: DEBUG oslo_concurrency.lockutils [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.090278] env[61768]: DEBUG oslo_concurrency.lockutils [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquired lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.090465] env[61768]: DEBUG nova.network.neutron [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Building network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1098.877807] env[61768]: DEBUG nova.network.neutron [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.315189] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.315479] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.380258] env[61768]: DEBUG oslo_concurrency.lockutils [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Releasing lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.380961] env[61768]: DEBUG nova.objects.instance [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lazy-loading 'flavor' on Instance uuid c4a4d1d5-d91f-4513-9dff-c0a5398ead5e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1099.819522] env[61768]: DEBUG nova.compute.utils [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Using /dev/sd instead of None {{(pid=61768) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1099.886852] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a74b7c-2d04-4535-b0b6-3a0eec936923 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.912254] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1099.912764] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54606497-f011-4317-9623-5d63d1f1ec41 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.920101] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1099.920101] env[61768]: value = "task-1229507" [ 1099.920101] env[61768]: _type = "Task" [ 1099.920101] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.928448] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229507, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.322497] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.430423] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229507, 'name': PowerOffVM_Task, 'duration_secs': 0.184146} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.430739] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1100.436045] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfiguring VM instance instance-0000006e to detach disk 2002 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1100.436329] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac7f3e54-cd1c-4123-9328-80ba04f102e5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.454047] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1100.454047] env[61768]: value = "task-1229508" [ 1100.454047] env[61768]: _type = "Task" [ 1100.454047] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.461786] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229508, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.963206] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229508, 'name': ReconfigVM_Task, 'duration_secs': 0.264951} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.963479] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfigured VM instance instance-0000006e to detach disk 2002 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1100.963676] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powering on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1100.963968] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf1c0ee2-3713-4272-8846-93ecee8050ed {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.970340] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1100.970340] env[61768]: value = "task-1229509" [ 1100.970340] env[61768]: _type = "Task" [ 1100.970340] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.978915] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229509, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.379497] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.379882] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.379962] env[61768]: INFO nova.compute.manager [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Attaching volume 0c4afb88-f782-410c-ae87-f8988ded06af to /dev/sdb [ 1101.411793] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85048184-47dd-4034-855d-1f93039073e6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.419114] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e10658a-dd13-4c5b-a48e-3e8739d18037 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.432453] env[61768]: DEBUG nova.virt.block_device [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updating existing volume attachment record: 55084f7a-d576-45d2-b963-92e91f725d55 {{(pid=61768) _volume_attach /opt/stack/nova/nova/virt/block_device.py:679}} [ 1101.480662] env[61768]: DEBUG oslo_vmware.api [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229509, 'name': PowerOnVM_Task, 'duration_secs': 0.360079} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.480950] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powered on the VM {{(pid=61768) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1101.481127] env[61768]: DEBUG nova.compute.manager [None req-60702d15-03be-4ec3-9a53-5ab0e11107a7 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Checking state {{(pid=61768) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.481901] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db154e62-00d9-4ba2-8357-bfc92d85823c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.876464] env[61768]: DEBUG nova.compute.manager [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Received event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.876729] env[61768]: DEBUG nova.compute.manager [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing instance network info cache due to event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1102.876914] env[61768]: DEBUG oslo_concurrency.lockutils [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] Acquiring lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.877080] env[61768]: DEBUG oslo_concurrency.lockutils [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] Acquired lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.877255] env[61768]: DEBUG nova.network.neutron [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1103.601784] env[61768]: DEBUG nova.network.neutron [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updated VIF entry in instance network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1103.602366] env[61768]: DEBUG nova.network.neutron [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.105039] env[61768]: DEBUG oslo_concurrency.lockutils [req-9dd0b419-ad56-4ba7-a8eb-d26e939eb16d req-273c6345-b3b5-4f70-bbc4-61d444b6b20a service nova] Releasing lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1104.993681] env[61768]: DEBUG nova.compute.manager [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Received event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.993902] env[61768]: DEBUG nova.compute.manager [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing instance network info cache due to event network-changed-3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1104.994150] env[61768]: DEBUG oslo_concurrency.lockutils [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] Acquiring lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.994308] env[61768]: DEBUG oslo_concurrency.lockutils [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] Acquired lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.994478] env[61768]: DEBUG nova.network.neutron [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Refreshing network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1105.695648] env[61768]: DEBUG nova.network.neutron [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updated VIF entry in instance network info cache for port 3770838d-f101-43db-99f9-fc4b33c5bda7. {{(pid=61768) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1105.696045] env[61768]: DEBUG nova.network.neutron [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [{"id": "3770838d-f101-43db-99f9-fc4b33c5bda7", "address": "fa:16:3e:dd:81:67", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3770838d-f1", "ovs_interfaceid": "3770838d-f101-43db-99f9-fc4b33c5bda7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.979716] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Volume attach. Driver type: vmdk {{(pid=61768) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1105.980032] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265551', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'name': 'volume-0c4afb88-f782-410c-ae87-f8988ded06af', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4d34dea4-917e-41a3-ac7c-706e97603711', 'attached_at': '', 'detached_at': '', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'serial': '0c4afb88-f782-410c-ae87-f8988ded06af'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1105.980947] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3829d125-d5a7-41dc-af3a-fc2e9c777ca5 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.641452] env[61768]: DEBUG oslo_concurrency.lockutils [req-f8edce8e-b879-45ed-a572-d7851e17c26e req-00f03fd7-0255-4c96-afb5-ed2f5f75a51c service nova] Releasing lock "refresh_cache-c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.642785] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f9ed39-0271-4441-8105-5ff0a199ce50 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.667471] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] volume-0c4afb88-f782-410c-ae87-f8988ded06af/volume-0c4afb88-f782-410c-ae87-f8988ded06af.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.667727] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b390328c-78e6-4e84-bdcd-14d5d17057a3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.685947] env[61768]: DEBUG oslo_vmware.api [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1106.685947] env[61768]: value = "task-1229512" [ 1106.685947] env[61768]: _type = "Task" [ 1106.685947] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.693957] env[61768]: DEBUG oslo_vmware.api [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229512, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.195494] env[61768]: DEBUG oslo_vmware.api [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229512, 'name': ReconfigVM_Task, 'duration_secs': 0.367774} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.195868] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Reconfigured VM instance instance-0000006f to attach disk [datastore1] volume-0c4afb88-f782-410c-ae87-f8988ded06af/volume-0c4afb88-f782-410c-ae87-f8988ded06af.vmdk or device None with type thin {{(pid=61768) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1107.200329] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b6a1f40a-b921-4896-b898-4b50479875e7 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.215394] env[61768]: DEBUG oslo_vmware.api [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1107.215394] env[61768]: value = "task-1229513" [ 1107.215394] env[61768]: _type = "Task" [ 1107.215394] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.222824] env[61768]: DEBUG oslo_vmware.api [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229513, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.724829] env[61768]: DEBUG oslo_vmware.api [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229513, 'name': ReconfigVM_Task, 'duration_secs': 0.137014} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.725141] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265551', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'name': 'volume-0c4afb88-f782-410c-ae87-f8988ded06af', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4d34dea4-917e-41a3-ac7c-706e97603711', 'attached_at': '', 'detached_at': '', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'serial': '0c4afb88-f782-410c-ae87-f8988ded06af'} {{(pid=61768) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1108.762653] env[61768]: DEBUG nova.objects.instance [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'flavor' on Instance uuid 4d34dea4-917e-41a3-ac7c-706e97603711 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1109.267891] env[61768]: DEBUG oslo_concurrency.lockutils [None req-0a330c80-49a4-48f8-bc93-095c773d9433 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.888s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.830593] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1109.831093] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.335350] env[61768]: INFO nova.compute.manager [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Detaching volume 0c4afb88-f782-410c-ae87-f8988ded06af [ 1110.366321] env[61768]: INFO nova.virt.block_device [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Attempting to driver detach volume 0c4afb88-f782-410c-ae87-f8988ded06af from mountpoint /dev/sdb [ 1110.366557] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1110.366742] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265551', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'name': 'volume-0c4afb88-f782-410c-ae87-f8988ded06af', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4d34dea4-917e-41a3-ac7c-706e97603711', 'attached_at': '', 'detached_at': '', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'serial': '0c4afb88-f782-410c-ae87-f8988ded06af'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1110.367680] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd76cb14-964b-41f8-9425-8c9590c96641 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.388544] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cbb776-2580-428e-9ebd-2c2e002e6519 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.395143] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959a1fea-e30d-454d-a3e8-4a4269ae6496 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.414767] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60fabb5b-2de2-4705-883a-6c4af162f9a9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.428299] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] The volume has not been displaced from its original location: [datastore1] volume-0c4afb88-f782-410c-ae87-f8988ded06af/volume-0c4afb88-f782-410c-ae87-f8988ded06af.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1110.433386] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1110.433632] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2ef0aed-76bb-412a-8a05-15966dc7130c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.451153] env[61768]: DEBUG oslo_vmware.api [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1110.451153] env[61768]: value = "task-1229514" [ 1110.451153] env[61768]: _type = "Task" [ 1110.451153] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.458442] env[61768]: DEBUG oslo_vmware.api [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229514, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.923353] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.923762] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.923762] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1110.923911] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1110.960640] env[61768]: DEBUG oslo_vmware.api [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229514, 'name': ReconfigVM_Task, 'duration_secs': 0.211665} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.960915] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1110.966194] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd399b5e-d297-4a3a-b775-1981a3f9a48d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.981365] env[61768]: DEBUG oslo_vmware.api [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1110.981365] env[61768]: value = "task-1229515" [ 1110.981365] env[61768]: _type = "Task" [ 1110.981365] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.988836] env[61768]: DEBUG oslo_vmware.api [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229515, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.427021] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.427278] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.427460] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1111.427618] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1111.428521] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aed494-67c8-4f96-988b-0772866c6896 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.437045] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba8ba0c-606a-4ee1-85cf-ecdf1d4af6e8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.451533] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f7c9812-ac01-4977-9cf3-204460f8c769 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.457533] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76d5fadb-c6c2-4775-a66b-16a390e27a8b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.485839] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180425MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1111.485996] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1111.486179] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1111.495051] env[61768]: DEBUG oslo_vmware.api [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229515, 'name': ReconfigVM_Task, 'duration_secs': 0.132981} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.495336] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265551', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'name': 'volume-0c4afb88-f782-410c-ae87-f8988ded06af', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '4d34dea4-917e-41a3-ac7c-706e97603711', 'attached_at': '', 'detached_at': '', 'volume_id': '0c4afb88-f782-410c-ae87-f8988ded06af', 'serial': '0c4afb88-f782-410c-ae87-f8988ded06af'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1112.038557] env[61768]: DEBUG nova.objects.instance [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'flavor' on Instance uuid 4d34dea4-917e-41a3-ac7c-706e97603711 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.515295] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance e8095d6d-18c5-4ca9-8ac8-475aeeffc01a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1112.515462] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance aa568a72-afba-4a25-a048-ca62191f9212 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1112.515585] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance c4a4d1d5-d91f-4513-9dff-c0a5398ead5e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1112.515704] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Instance 4d34dea4-917e-41a3-ac7c-706e97603711 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61768) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1112.515884] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1112.516075] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1112.570770] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f844b93-7ecb-4ad9-a148-5ce799374742 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.577791] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0587d51-f821-4022-8617-e1dc9b3e4d53 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.607475] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d33421-7a63-493c-a370-f8d973f9466f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.614017] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26c5662-1fe6-4b05-8c80-0a021bf069fa {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.626479] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.045584] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c51ac939-59a7-4bd9-a70f-ba63a3bcc4ed tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.214s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.129484] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.634424] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1113.634666] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.148s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.062416] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.062817] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.062881] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "4d34dea4-917e-41a3-ac7c-706e97603711-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.063051] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.063234] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.065497] env[61768]: INFO nova.compute.manager [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Terminating instance [ 1114.067271] env[61768]: DEBUG nova.compute.manager [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1114.067469] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1114.068313] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db4b8dc3-4eb7-4960-a43d-8509cd46ca03 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.076037] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1114.076282] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f5b56756-8bb9-47d8-a7de-887ffc0e8110 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.082580] env[61768]: DEBUG oslo_vmware.api [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1114.082580] env[61768]: value = "task-1229516" [ 1114.082580] env[61768]: _type = "Task" [ 1114.082580] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.090708] env[61768]: DEBUG oslo_vmware.api [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229516, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.592312] env[61768]: DEBUG oslo_vmware.api [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229516, 'name': PowerOffVM_Task, 'duration_secs': 0.158985} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.592559] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1114.592731] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1114.592968] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a70b373a-9f59-44a1-a0b5-5f8cf29be639 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.656162] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1114.656441] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1114.656648] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleting the datastore file [datastore2] 4d34dea4-917e-41a3-ac7c-706e97603711 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1114.656906] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-88b6789e-1880-4936-950c-bde05f658dc9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.662654] env[61768]: DEBUG oslo_vmware.api [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for the task: (returnval){ [ 1114.662654] env[61768]: value = "task-1229518" [ 1114.662654] env[61768]: _type = "Task" [ 1114.662654] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.670166] env[61768]: DEBUG oslo_vmware.api [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229518, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.173144] env[61768]: DEBUG oslo_vmware.api [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Task: {'id': task-1229518, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146434} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.173536] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1115.173608] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1115.173790] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1115.173971] env[61768]: INFO nova.compute.manager [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1115.174249] env[61768]: DEBUG oslo.service.loopingcall [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1115.174472] env[61768]: DEBUG nova.compute.manager [-] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1115.174570] env[61768]: DEBUG nova.network.neutron [-] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1115.325614] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1115.325788] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1115.585435] env[61768]: DEBUG nova.compute.manager [req-099a106c-9a7d-4820-8adf-fa5fdab4afb1 req-cf4971bf-d783-408f-a497-eaa243275b62 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Received event network-vif-deleted-8ec9169c-da5b-4f41-9bd9-76d42ac011e6 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1115.585678] env[61768]: INFO nova.compute.manager [req-099a106c-9a7d-4820-8adf-fa5fdab4afb1 req-cf4971bf-d783-408f-a497-eaa243275b62 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Neutron deleted interface 8ec9169c-da5b-4f41-9bd9-76d42ac011e6; detaching it from the instance and deleting it from the info cache [ 1115.585820] env[61768]: DEBUG nova.network.neutron [req-099a106c-9a7d-4820-8adf-fa5fdab4afb1 req-cf4971bf-d783-408f-a497-eaa243275b62 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.856598] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1115.856769] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquired lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1115.856918] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Forcefully refreshing network info cache for instance {{(pid=61768) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1116.063187] env[61768]: DEBUG nova.network.neutron [-] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.088401] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dcac1259-b8a3-473e-a9f8-7e8dbccd830d {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.099124] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17630ac6-8be8-4143-a2b3-b519284a386b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.125090] env[61768]: DEBUG nova.compute.manager [req-099a106c-9a7d-4820-8adf-fa5fdab4afb1 req-cf4971bf-d783-408f-a497-eaa243275b62 service nova] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Detach interface failed, port_id=8ec9169c-da5b-4f41-9bd9-76d42ac011e6, reason: Instance 4d34dea4-917e-41a3-ac7c-706e97603711 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1116.565905] env[61768]: INFO nova.compute.manager [-] [instance: 4d34dea4-917e-41a3-ac7c-706e97603711] Took 1.39 seconds to deallocate network for instance. [ 1117.059366] env[61768]: DEBUG nova.network.neutron [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updating instance_info_cache with network_info: [{"id": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "address": "fa:16:3e:b3:4f:18", "network": {"id": "d0e4c464-6402-4a38-ae93-ce5d1c798be2", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1455858200-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e4c2ff63d9d94a41b0dbc7d597e762a8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4162774e-ec80-4d85-aeb4-fae77f197393", "external-id": "nsx-vlan-transportzone-542", "segmentation_id": 542, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb4f2467a-e3", "ovs_interfaceid": "b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.072278] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.072754] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.073012] env[61768]: DEBUG nova.objects.instance [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lazy-loading 'resources' on Instance uuid 4d34dea4-917e-41a3-ac7c-706e97603711 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.561915] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Releasing lock "refresh_cache-e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" {{(pid=61768) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1117.562181] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updated the network info_cache for instance {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1117.562405] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.562571] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.562728] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.562875] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1117.633617] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3e11d9-6cba-40f1-8ef4-595057716b39 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.641043] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa2f01a9-1e7a-4fc1-b18b-822f0fba016f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.670539] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-105c92d5-d90b-4a90-836d-37d2ec7c6691 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.677373] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4831b4f4-4bcf-48b0-9d70-0d5590365c5a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.689659] env[61768]: DEBUG nova.compute.provider_tree [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1118.192986] env[61768]: DEBUG nova.scheduler.client.report [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.698507] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.626s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.718632] env[61768]: INFO nova.scheduler.client.report [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Deleted allocations for instance 4d34dea4-917e-41a3-ac7c-706e97603711 [ 1118.847374] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1119.227372] env[61768]: DEBUG oslo_concurrency.lockutils [None req-e34ecedf-3642-4eef-bc2d-476f86c6d447 tempest-AttachVolumeNegativeTest-1226778607 tempest-AttachVolumeNegativeTest-1226778607-project-member] Lock "4d34dea4-917e-41a3-ac7c-706e97603711" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.165s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.266378] env[61768]: DEBUG oslo_concurrency.lockutils [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.266720] env[61768]: DEBUG oslo_concurrency.lockutils [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.770529] env[61768]: INFO nova.compute.manager [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Detaching volume cbb7e026-7118-4822-b03a-1afcd8b91b04 [ 1138.801281] env[61768]: INFO nova.virt.block_device [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Attempting to driver detach volume cbb7e026-7118-4822-b03a-1afcd8b91b04 from mountpoint /dev/sdb [ 1138.801534] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Volume detach. Driver type: vmdk {{(pid=61768) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1138.801922] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265550', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'name': 'volume-cbb7e026-7118-4822-b03a-1afcd8b91b04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4a4d1d5-d91f-4513-9dff-c0a5398ead5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'serial': 'cbb7e026-7118-4822-b03a-1afcd8b91b04'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1138.802681] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a104d993-33ff-462c-b688-c1e0ca1533b6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.823920] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d0c9ccf-9867-4ad2-b2cf-6b4c6912646f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.830277] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e92053-b498-46f6-8836-2bed87574d37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.849201] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f41ad25-f4e4-43f4-a5da-39ba5bd626eb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.862940] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] The volume has not been displaced from its original location: [datastore2] volume-cbb7e026-7118-4822-b03a-1afcd8b91b04/volume-cbb7e026-7118-4822-b03a-1afcd8b91b04.vmdk. No consolidation needed. {{(pid=61768) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1138.868124] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfiguring VM instance instance-0000006e to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1138.868379] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-babde54f-c59a-4bc5-845c-a0863959ec62 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.885487] env[61768]: DEBUG oslo_vmware.api [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1138.885487] env[61768]: value = "task-1229522" [ 1138.885487] env[61768]: _type = "Task" [ 1138.885487] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1138.892634] env[61768]: DEBUG oslo_vmware.api [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229522, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.394815] env[61768]: DEBUG oslo_vmware.api [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229522, 'name': ReconfigVM_Task, 'duration_secs': 0.223166} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.395254] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Reconfigured VM instance instance-0000006e to detach disk 2001 {{(pid=61768) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1139.399676] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d083a033-86c9-4971-88a9-6f5908b930f1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.414610] env[61768]: DEBUG oslo_vmware.api [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1139.414610] env[61768]: value = "task-1229523" [ 1139.414610] env[61768]: _type = "Task" [ 1139.414610] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.421972] env[61768]: DEBUG oslo_vmware.api [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229523, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.924254] env[61768]: DEBUG oslo_vmware.api [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229523, 'name': ReconfigVM_Task, 'duration_secs': 0.160462} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1139.924570] env[61768]: DEBUG nova.virt.vmwareapi.volumeops [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-265550', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'name': 'volume-cbb7e026-7118-4822-b03a-1afcd8b91b04', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c4a4d1d5-d91f-4513-9dff-c0a5398ead5e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cbb7e026-7118-4822-b03a-1afcd8b91b04', 'serial': 'cbb7e026-7118-4822-b03a-1afcd8b91b04'} {{(pid=61768) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1140.464676] env[61768]: DEBUG nova.objects.instance [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lazy-loading 'flavor' on Instance uuid c4a4d1d5-d91f-4513-9dff-c0a5398ead5e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1141.473678] env[61768]: DEBUG oslo_concurrency.lockutils [None req-33fc4965-9cc5-4318-97f6-a4b26af3accb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.207s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.557102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.557102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.557102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.557102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.557102] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.559348] env[61768]: INFO nova.compute.manager [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Terminating instance [ 1142.561235] env[61768]: DEBUG nova.compute.manager [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1142.561431] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1142.562280] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8401fd47-e963-4bc0-bc6a-cdab10184bfd {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.570469] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1142.570698] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6a40950e-1a8a-4f09-a7b6-13d81fb102bc {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.577015] env[61768]: DEBUG oslo_vmware.api [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1142.577015] env[61768]: value = "task-1229524" [ 1142.577015] env[61768]: _type = "Task" [ 1142.577015] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.584364] env[61768]: DEBUG oslo_vmware.api [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229524, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.086943] env[61768]: DEBUG oslo_vmware.api [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229524, 'name': PowerOffVM_Task, 'duration_secs': 0.167584} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.087230] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1143.087407] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1143.087654] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-49569b1a-335f-44f5-b2e3-2053d3f77468 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.147138] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1143.147420] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Deleting contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1143.147654] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleting the datastore file [datastore2] c4a4d1d5-d91f-4513-9dff-c0a5398ead5e {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1143.147930] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-96c6c69a-13f7-4594-80a3-24b6f28f573e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.153844] env[61768]: DEBUG oslo_vmware.api [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1143.153844] env[61768]: value = "task-1229526" [ 1143.153844] env[61768]: _type = "Task" [ 1143.153844] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1143.161173] env[61768]: DEBUG oslo_vmware.api [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.663544] env[61768]: DEBUG oslo_vmware.api [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.162618} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.663819] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1143.663993] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Deleted contents of the VM from datastore datastore2 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1143.664198] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1143.664382] env[61768]: INFO nova.compute.manager [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1143.664640] env[61768]: DEBUG oslo.service.loopingcall [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1143.664834] env[61768]: DEBUG nova.compute.manager [-] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1143.664930] env[61768]: DEBUG nova.network.neutron [-] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1144.106281] env[61768]: DEBUG nova.compute.manager [req-47d51f5c-bfff-4216-b4ea-796bdfa1e55e req-7001ba72-77a1-4761-9aa8-bbfaf316a606 service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Received event network-vif-deleted-3770838d-f101-43db-99f9-fc4b33c5bda7 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1144.106488] env[61768]: INFO nova.compute.manager [req-47d51f5c-bfff-4216-b4ea-796bdfa1e55e req-7001ba72-77a1-4761-9aa8-bbfaf316a606 service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Neutron deleted interface 3770838d-f101-43db-99f9-fc4b33c5bda7; detaching it from the instance and deleting it from the info cache [ 1144.106659] env[61768]: DEBUG nova.network.neutron [req-47d51f5c-bfff-4216-b4ea-796bdfa1e55e req-7001ba72-77a1-4761-9aa8-bbfaf316a606 service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.581623] env[61768]: DEBUG nova.network.neutron [-] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.609975] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-38798dc3-430e-4d15-9560-75dcb639ac43 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.619971] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4526e9a8-5082-4928-b70d-76b79599133f {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.644866] env[61768]: DEBUG nova.compute.manager [req-47d51f5c-bfff-4216-b4ea-796bdfa1e55e req-7001ba72-77a1-4761-9aa8-bbfaf316a606 service nova] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Detach interface failed, port_id=3770838d-f101-43db-99f9-fc4b33c5bda7, reason: Instance c4a4d1d5-d91f-4513-9dff-c0a5398ead5e could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1145.084554] env[61768]: INFO nova.compute.manager [-] [instance: c4a4d1d5-d91f-4513-9dff-c0a5398ead5e] Took 1.42 seconds to deallocate network for instance. [ 1145.591760] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1145.591995] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.592242] env[61768]: DEBUG nova.objects.instance [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lazy-loading 'resources' on Instance uuid c4a4d1d5-d91f-4513-9dff-c0a5398ead5e {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1146.147223] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd44ee1-4010-4fec-9df7-f5da53696e16 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.154691] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98b4e1ed-653a-4642-ba6f-eb6e4cb5a89c {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.183416] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca6b3c77-9487-41a6-ae83-28669f91d674 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.189864] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fecc81e-2b0f-47b9-a99b-aff70b8ebde2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.202372] env[61768]: DEBUG nova.compute.provider_tree [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1146.705169] env[61768]: DEBUG nova.scheduler.client.report [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1147.211316] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.619s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.235399] env[61768]: INFO nova.scheduler.client.report [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleted allocations for instance c4a4d1d5-d91f-4513-9dff-c0a5398ead5e [ 1147.745454] env[61768]: DEBUG oslo_concurrency.lockutils [None req-099ed6fc-c1f5-41e2-898f-0b276b2859cb tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "c4a4d1d5-d91f-4513-9dff-c0a5398ead5e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.188s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.630979] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "aa568a72-afba-4a25-a048-ca62191f9212" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.631374] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "aa568a72-afba-4a25-a048-ca62191f9212" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.631451] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "aa568a72-afba-4a25-a048-ca62191f9212-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1152.631672] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "aa568a72-afba-4a25-a048-ca62191f9212-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.631865] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "aa568a72-afba-4a25-a048-ca62191f9212-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1152.635136] env[61768]: INFO nova.compute.manager [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Terminating instance [ 1152.636841] env[61768]: DEBUG nova.compute.manager [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1152.637061] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1152.637924] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6902924-aa89-4b14-9866-4123538699f3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.646259] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1152.646486] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-07b33b05-7ed3-4a86-9efe-f72a0fa14ef4 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.652015] env[61768]: DEBUG oslo_vmware.api [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1152.652015] env[61768]: value = "task-1229528" [ 1152.652015] env[61768]: _type = "Task" [ 1152.652015] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.161950] env[61768]: DEBUG oslo_vmware.api [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229528, 'name': PowerOffVM_Task, 'duration_secs': 0.255945} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.162234] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1153.162444] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1153.162693] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-371b4712-9811-4a8e-b81e-271b529586eb {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.221764] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1153.221987] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1153.222196] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleting the datastore file [datastore1] aa568a72-afba-4a25-a048-ca62191f9212 {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1153.222500] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84186825-8f46-4a0d-be5d-7f598a71a998 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.228640] env[61768]: DEBUG oslo_vmware.api [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1153.228640] env[61768]: value = "task-1229530" [ 1153.228640] env[61768]: _type = "Task" [ 1153.228640] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.235662] env[61768]: DEBUG oslo_vmware.api [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229530, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.738327] env[61768]: DEBUG oslo_vmware.api [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229530, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158636} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.738750] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1153.738750] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1153.738907] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1153.739098] env[61768]: INFO nova.compute.manager [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1153.739350] env[61768]: DEBUG oslo.service.loopingcall [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1153.739546] env[61768]: DEBUG nova.compute.manager [-] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1153.739640] env[61768]: DEBUG nova.network.neutron [-] [instance: aa568a72-afba-4a25-a048-ca62191f9212] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1154.154085] env[61768]: DEBUG nova.compute.manager [req-c2a6a0f4-f57b-4847-bb29-5ed60eb0c7f6 req-9f52548b-684c-4577-9249-d4e4ddeb7760 service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Received event network-vif-deleted-5913b4b2-1038-4537-9cd9-b466dc48d869 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1154.154219] env[61768]: INFO nova.compute.manager [req-c2a6a0f4-f57b-4847-bb29-5ed60eb0c7f6 req-9f52548b-684c-4577-9249-d4e4ddeb7760 service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Neutron deleted interface 5913b4b2-1038-4537-9cd9-b466dc48d869; detaching it from the instance and deleting it from the info cache [ 1154.154404] env[61768]: DEBUG nova.network.neutron [req-c2a6a0f4-f57b-4847-bb29-5ed60eb0c7f6 req-9f52548b-684c-4577-9249-d4e4ddeb7760 service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.633783] env[61768]: DEBUG nova.network.neutron [-] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1154.656941] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6466caf0-bbcd-4eba-9b4d-859037c23295 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.666915] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-777558b2-1eaf-4d8b-b5e6-7bb15ee56012 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.688369] env[61768]: DEBUG nova.compute.manager [req-c2a6a0f4-f57b-4847-bb29-5ed60eb0c7f6 req-9f52548b-684c-4577-9249-d4e4ddeb7760 service nova] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Detach interface failed, port_id=5913b4b2-1038-4537-9cd9-b466dc48d869, reason: Instance aa568a72-afba-4a25-a048-ca62191f9212 could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1155.136648] env[61768]: INFO nova.compute.manager [-] [instance: aa568a72-afba-4a25-a048-ca62191f9212] Took 1.40 seconds to deallocate network for instance. [ 1155.643911] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1155.644195] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1155.644449] env[61768]: DEBUG nova.objects.instance [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lazy-loading 'resources' on Instance uuid aa568a72-afba-4a25-a048-ca62191f9212 {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1156.197103] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f74fdc0-fd9e-43bb-bbba-e3827a154cf8 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.204981] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f07426-a6fa-4b1d-9bb7-b05a2f69871a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.235096] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-551f8293-0dff-4383-98c6-0e84e9b5773e {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.242550] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff18c73-6b5e-4680-8e6e-2e2b538967e3 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.257719] env[61768]: DEBUG nova.compute.provider_tree [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1156.761394] env[61768]: DEBUG nova.scheduler.client.report [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1157.267256] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.288387] env[61768]: INFO nova.scheduler.client.report [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleted allocations for instance aa568a72-afba-4a25-a048-ca62191f9212 [ 1157.797934] env[61768]: DEBUG oslo_concurrency.lockutils [None req-2be397e6-57a8-4afa-ac2e-248fdb9d9259 tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "aa568a72-afba-4a25-a048-ca62191f9212" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.166s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.185273] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.185555] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.185773] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1158.185966] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.186167] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.188399] env[61768]: INFO nova.compute.manager [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Terminating instance [ 1158.190286] env[61768]: DEBUG nova.compute.manager [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Start destroying the instance on the hypervisor. {{(pid=61768) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1158.190490] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Destroying instance {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1158.191332] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e0d3dbc-0212-41c6-bffb-0de4e916a161 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.199239] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powering off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1158.199470] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-572778fd-94db-43d5-9343-2d34d2ca524b {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.206664] env[61768]: DEBUG oslo_vmware.api [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1158.206664] env[61768]: value = "task-1229531" [ 1158.206664] env[61768]: _type = "Task" [ 1158.206664] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.214716] env[61768]: DEBUG oslo_vmware.api [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.716051] env[61768]: DEBUG oslo_vmware.api [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229531, 'name': PowerOffVM_Task, 'duration_secs': 0.185162} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.716447] env[61768]: DEBUG nova.virt.vmwareapi.vm_util [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Powered off the VM {{(pid=61768) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1158.716551] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Unregistering the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1158.716835] env[61768]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc7dbe74-57f3-462f-b937-6f3111fadb11 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.778747] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Unregistered the VM {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1158.778989] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Deleting contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1158.779189] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleting the datastore file [datastore1] e8095d6d-18c5-4ca9-8ac8-475aeeffc01a {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.779453] env[61768]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5b9cd394-152f-4dc0-a97e-31c2553b9f01 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.785634] env[61768]: DEBUG oslo_vmware.api [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for the task: (returnval){ [ 1158.785634] env[61768]: value = "task-1229533" [ 1158.785634] env[61768]: _type = "Task" [ 1158.785634] env[61768]: } to complete. {{(pid=61768) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.793041] env[61768]: DEBUG oslo_vmware.api [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229533, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.295754] env[61768]: DEBUG oslo_vmware.api [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Task: {'id': task-1229533, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139952} completed successfully. {{(pid=61768) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.295978] env[61768]: DEBUG nova.virt.vmwareapi.ds_util [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleted the datastore file {{(pid=61768) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1159.296210] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Deleted contents of the VM from datastore datastore1 {{(pid=61768) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1159.296370] env[61768]: DEBUG nova.virt.vmwareapi.vmops [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Instance destroyed {{(pid=61768) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1159.296551] env[61768]: INFO nova.compute.manager [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1159.296800] env[61768]: DEBUG oslo.service.loopingcall [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61768) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1159.296996] env[61768]: DEBUG nova.compute.manager [-] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Deallocating network for instance {{(pid=61768) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1159.297107] env[61768]: DEBUG nova.network.neutron [-] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] deallocate_for_instance() {{(pid=61768) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1159.522954] env[61768]: DEBUG nova.compute.manager [req-b84f0537-d762-49bd-bb19-953e09176665 req-81879cea-6c13-4f7f-8f50-6bbfdb7b8aee service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Received event network-vif-deleted-b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8 {{(pid=61768) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1159.523218] env[61768]: INFO nova.compute.manager [req-b84f0537-d762-49bd-bb19-953e09176665 req-81879cea-6c13-4f7f-8f50-6bbfdb7b8aee service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Neutron deleted interface b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8; detaching it from the instance and deleting it from the info cache [ 1159.523366] env[61768]: DEBUG nova.network.neutron [req-b84f0537-d762-49bd-bb19-953e09176665 req-81879cea-6c13-4f7f-8f50-6bbfdb7b8aee service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.003603] env[61768]: DEBUG nova.network.neutron [-] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Updating instance_info_cache with network_info: [] {{(pid=61768) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1160.026255] env[61768]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-045f4b4e-dbd3-4cbe-846a-c290c50683af {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.035314] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ede9a7-e0b7-420e-b054-7c548a8a6dd2 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.057042] env[61768]: DEBUG nova.compute.manager [req-b84f0537-d762-49bd-bb19-953e09176665 req-81879cea-6c13-4f7f-8f50-6bbfdb7b8aee service nova] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Detach interface failed, port_id=b4f2467a-e3ee-40b1-9fb9-ab9499a7ccd8, reason: Instance e8095d6d-18c5-4ca9-8ac8-475aeeffc01a could not be found. {{(pid=61768) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1160.506764] env[61768]: INFO nova.compute.manager [-] [instance: e8095d6d-18c5-4ca9-8ac8-475aeeffc01a] Took 1.21 seconds to deallocate network for instance. [ 1161.013690] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1161.013979] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1161.014215] env[61768]: DEBUG nova.objects.instance [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lazy-loading 'resources' on Instance uuid e8095d6d-18c5-4ca9-8ac8-475aeeffc01a {{(pid=61768) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.547597] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec509355-367b-4253-a4ca-40731db4010a {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.555327] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-556bede4-c475-42d6-8a2a-1576b1f6e148 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.584056] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd53dbfe-0692-419f-96d9-9a3921d09d37 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.590627] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712a1bb1-9bda-4a0d-b0aa-d85ba1791ba6 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.604254] env[61768]: DEBUG nova.compute.provider_tree [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1162.107590] env[61768]: DEBUG nova.scheduler.client.report [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.612160] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.598s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.629694] env[61768]: INFO nova.scheduler.client.report [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Deleted allocations for instance e8095d6d-18c5-4ca9-8ac8-475aeeffc01a [ 1163.137303] env[61768]: DEBUG oslo_concurrency.lockutils [None req-c86062f3-e51d-444b-b479-69a648f06ece tempest-ServerRescueNegativeTestJSON-656530817 tempest-ServerRescueNegativeTestJSON-656530817-project-member] Lock "e8095d6d-18c5-4ca9-8ac8-475aeeffc01a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.952s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.615608] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager.update_available_resource {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1169.118717] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.118967] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.119167] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.119327] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61768) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1169.120253] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-844eb4ad-6b85-4572-a045-ee09dec45e67 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.128237] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a450801-d492-4e12-ad50-c69e407e87c1 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.141533] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678f0068-277f-4dff-95cd-285d53c4d1f9 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.147419] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7003727c-dd58-4d86-a94e-d0c06d247b66 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.176919] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180651MB free_disk=151GB free_vcpus=48 pci_devices=None {{(pid=61768) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1169.177072] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.177298] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1170.196635] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1170.196855] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61768) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1170.209086] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b661922-03c3-4ee9-b9b4-4a7761f288da {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.216658] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04c12a9-4cd9-4309-b024-32126f44a138 {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.244968] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfdbbee-bfad-4943-9097-50c7092c4cde {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.251343] env[61768]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22bf20e9-822c-4f06-aa8b-f7a582c0d1ad {{(pid=61768) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.263520] env[61768]: DEBUG nova.compute.provider_tree [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed in ProviderTree for provider: f9f7149c-7c4e-44e7-8000-07a202159d2f {{(pid=61768) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.766306] env[61768]: DEBUG nova.scheduler.client.report [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Inventory has not changed for provider f9f7149c-7c4e-44e7-8000-07a202159d2f based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 151, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61768) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.271599] env[61768]: DEBUG nova.compute.resource_tracker [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61768) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1171.271960] env[61768]: DEBUG oslo_concurrency.lockutils [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.095s {{(pid=61768) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1173.271108] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1173.271466] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1173.271466] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61768) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1173.615520] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1174.615485] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1175.615109] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1175.615500] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Starting heal instance info cache {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1177.123284] env[61768]: DEBUG nova.compute.manager [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Didn't find any instances for network info cache update. {{(pid=61768) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1177.123659] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1177.615275] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1177.615521] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1179.610646] env[61768]: DEBUG oslo_service.periodic_task [None req-628a7fb6-1b7c-4736-aaa1-873d8385fc0f None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61768) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}}